WorldWideScience

Sample records for model promet-v process

  1. 2-way coupling the hydrological land surface model PROMET with the regional climate model MM5

    Directory of Open Access Journals (Sweden)

    F. Zabel

    2013-05-01

    Full Text Available Most land surface hydrological models (LSHMs consider land surface processes (e.g. soil–plant–atmosphere interactions, lateral water flows, snow and ice in a spatially detailed manner. The atmosphere is considered as exogenous driver, neglecting feedbacks between the land surface and the atmosphere. On the other hand, regional climate models (RCMs generally simulate land surface processes through coarse descriptions and spatial scales but include land–atmosphere interactions. What is the impact of the differently applied model physics and spatial resolution of LSHMs on the performance of RCMs? What feedback effects are induced by different land surface models? This study analyses the impact of replacing the land surface module (LSM within an RCM with a high resolution LSHM. A 2-way coupling approach was applied using the LSHM PROMET (1 × 1 km2 and the atmospheric part of the RCM MM5 (45 × 45 km2. The scaling interface SCALMET is used for down- and upscaling the linear and non-linear fluxes between the model scales. The change in the atmospheric response by MM5 using the LSHM is analysed, and its quality is compared to observations of temperature and precipitation for a 4 yr period from 1996 to 1999 for the Upper Danube catchment. By substituting the Noah-LSM with PROMET, simulated non-bias-corrected near-surface air temperature improves for annual, monthly and daily courses when compared to measurements from 277 meteorological weather stations within the Upper Danube catchment. The mean annual bias was improved from −0.85 to −0.13 K. In particular, the improved afternoon heating from May to September is caused by increased sensible heat flux and decreased latent heat flux as well as more incoming solar radiation in the fully coupled PROMET/MM5 in comparison to the NOAH/MM5 simulation. Triggered by the LSM replacement, precipitation overall is reduced; however simulated precipitation amounts are still of high uncertainty, both

  2. PROMET - The Journal of Meteorological Education issued by DWD

    Science.gov (United States)

    Rapp, J.

    2009-09-01

    Promet is published by the German Meteorological Service (DWD) since 1971 to improve meteorologists and weather forecasters skills. The journal comprises mainly contributions to topics like biometeorology, the NAO, or meteorology and insurance business. The science-based articles should illustrate the special issue in an understandable and transparent way. In addition, the journal contains portraits of other national meteorological services and university departments, book reviews, list of university degrees, and other individual papers. Promet is published only in German language, but included English titles and abstracts. The journal is peer-reviewed by renowned external scientists. It is distributed free of charge by DWD to the own meteorological staff. On the other hand, DMG (the German Meteorological Society) hand it out to all members of the society. The current issues deal with "Modern procedures of weather forecasting in DWD” and "E-Learning in Meteorology”.

  3. Spatially Distributed, Coupled Modeling of Plant Growth, Nitrogen and Water Fluxes in an Alpine Catchment

    Science.gov (United States)

    Schneider, K.

    2001-12-01

    Carbon, water and nitrogen fluxes are closely coupled. They interact and have many feedbacks. Human interference, in particular through land use management and global change strongly modifies these fluxes. Increasing demands and conflicting interests result in an increasing need for regulation targeting different aspects of the system. Without being their main target, many of these measures directly affect water quantity, quality and availability. Improved management and planning of our water resources requires the development of integrated tools, in particular since interactions of the involved environmental and social systems often lead to unexpected or adverse results. To investigate the effect of plant growth, land use management and global change on water fluxes and quality, the PROcess oriented Modular EnvironmenT and Vegetation Model (PROMET-V) was developed. PROMET-V models the spatial patterns and temporal course of water, carbon and nitrogen fluxes using process oriented and mechanistic model components. The hydrological model is based on the Penman-Monteith approach, it uses a plant-physiological model to calculate the canopy conductance, and a multi-layer soil water model. Plant growth for different vegetation is modelled by calculating canopy photosynthesis, respiration, phenology and allocation. Plant growth and water fluxes are coupled directly through photosynthesis and transpiration. Many indirect feedbacks and interactions occur due to their mutual dependency upon leaf area, root distribution, water and nutrient availability for instance. PROMET-V calculates nitrogen fluxes and transformations. The time step used depends upon the modelled process and varies from 1 hour to 1 day. The kernel model is integrated in a raster GIS system for spatially distributed modelling. PROMET-V was tested in a pre-alpine landscape (Ammer river, 709 km**2, located in Southern Germany) which is characterized by small scale spatial heterogeneities of climate, soil and

  4. Povezovanje logov, logotipov in sloganov v zaščitnih grafičnih znakih

    OpenAIRE

    Ambrožič, Maja

    2016-01-01

    Ne glede na velikost organizacije, letni plačilni promet, organizacijsko strukturo, organizacijsko kulturo, naprave v lasti organizacije, poslovne prostore, organizacijsko klimo, organizacijsko strategijo, poslanstvo, vizijo, število zaposlenih itd v očeh javnosti vedno šteje oziroma organizacijo popolno predstavlja njena celostna grafična podoba. Celostna grafična podoba je identiteta organizacije, ki vključuje vso zgoraj našteto vsebino, ki jo mora grafični oblikovalec povezati v zaščitni g...

  5. Physics at a 100 TeV pp Collider: Standard Model Processes

    Energy Technology Data Exchange (ETDEWEB)

    Mangano, M. L. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Zanderighi, G. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Aguilar Saavedra, J. A. [Univ. of Granada (Spain); Alekhin, S. [Univ. of Hamburg (Germany). Inst. for Theoretical Physics; Inst. for High Energy Physics (IHEP), Moscow (Russian Federation); Badger, S. [Univ. of Edinburgh, Scotland (United Kingdom); Bauer, C. W. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Becher, T. [Univ. Bern (Switzerland); Bertone, V. [Univ. of Oxford (United Kingdom); Bonvini, M. [Univ. of Oxford (United Kingdom); Boselli, S. [Univ. of Pavia (Italy); Bothmann, E. [Gottingen Univ. (Germany); Boughezal, R. [Argonne National Lab. (ANL), Argonne, IL (United States); Cacciari, M. [Univ. Paris Diderot (France); Sorbonne Univ., Paris (France); Carloni Calame, C M. [Istituto Nazionale di Fisica Nucleare (INFN), Pavia (Italy); Caola, F. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Campbell, J. M. [Fermi National Accelerator Lab. (FNAL), Batavia, IL (United States); Carrazza, S. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Chiesa, M. [Istituto Nazionale di Fisica Nucleare (INFN), Pavia (Italy); Cieri, L. [Univ. of Zurich (Switzerland); Cimaglia, F. [Univ. degli Studi di Milano (Italy); Febres Cordero, F. [Physikalisches Inst., Freiburg (Germany); Ferrarese, P. [Gottingen Univ. (Germany); D' Enterria, D. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Ferrera, G. [Univ. degli Studi di Milano (Italy); Garcia i Tormo, X. [Univ. Bern (Switzerland); Garzelli, M. V. [Univ. of Hamburg (Germany); Germann, E. [Monash Univ., Melbourne, VIC (Australia); Hirschi, V. [SLAC National Accelerator Lab., Menlo Park, CA (United States); Han, T. [Univ. of Pittsburgh, PA (United States); Ita, H. [Physikalisches Inst., Freiburg (Germany); Jager, B. [Univ. of Tubingen (Germany); Kallweit, S. [Johannes Gutenberg Univ., Mainz (Germany); Karlberg, A. [Univ. of Oxford (United Kingdom); Kuttimalai, S. [Durham Univ. (United Kingdom); Krauss, F. [Durham Univ. (United Kingdom); Larkoski, A. J. [Harvard Univ., Cambridge, MA (United States); Lindert, J. [Univ. of Zurich (Switzerland); Luisoni, G. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Maierhofer, P. [Univ. of Freiburg (Germany); Mattelaer, O. [Durham Univ. (United Kingdom); Martinez, H. [Univ. of Pavia (Italy); Moch, S. [Univ. of Hamburg (Germany); Montagna, G. [Univ. of Pavia (Italy); Moretti, M. [Univ. of Ferrara (Italy); Nason, P. [Univ. of Milano (Italy); Nicrosini, O. [Istituto Nazionale di Fisica Nucleare (INFN), Pavia (Italy); Oleari, C. [Univ. of Milano (Italy); Pagani, D. [Univ. Catholique de Louvain (Belgium); Papaefstathiou, A. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Petriello, F. [Northwestern Univ., Evanston, IL (United States); Piccinini, F. [Istituto Nazionale di Fisica Nucleare (INFN), Pavia (Italy); Pierini, M. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Pierog, T. [Karlsruhe Inst. of Technology (KIT) (Germany); Pozzorini, S. [Univ. of Zurich (Switzerland); Re, E. [National Centre for Scientific Research (CNRS), Annecy-le-Vieux (France). Lab. of Annecy-le-Vieux for Theoretical Physics (LAPTh); Robens, T. [Technische Universitat Dresden (Germany); Rojo, J. [Univ. of Oxford (United Kingdom); Ruiz, R. [Durham Univ. (United Kingdom); Sakurai, K. [Durham Univ. (United Kingdom); Salam, G. P. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Salfelder, L. [Univ. of Tubingen (Germany); Schonherr, M. [Univ. of Ferrara (Italy); Schulze, M. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Schumann, S. [Univ. Gottingen (Germany); Selvaggi, M. [Univ. Catholique de Louvain (Belgium); Shivaji, A. [Istituto Nazionale di Fisica Nucleare (INFN), Pavia (Italy); Siodmok, A. [European Organization for Nuclear Research (CERN), Geneva (Switzerland); Polish Academy of Sciences (PAS), Krakow (Poland); Skands, P. [Monash Univ., Melbourne, VIC (Australia); Torrielli, P. [Univ. of Torino (Italy); Tramontano, F. [Univ. of Napoli (Italy); Tsinikos, I. [Univ. Catholique de Louvain (Belgium); Tweedie, B. [Univ. of Pittsburgh, PA (United States); Vicini, A. [Univ. degli Studi di Milano (Italy); Westhoff, S. [Heidelberg Univ. (Germany); Zaro, M. [Sorbonne Univ., Paris (France); Zeppenfeld, D. [Forschungszentrum Karlsruhe (Germany)

    2017-06-22

    This report summarises the properties of Standard Model processes at the 100 TeV pp collider. We document the production rates and typical distributions for a number of benchmark Standard Model processes, and discuss new dynamical phenomena arising at the highest energies available at this collider. We discuss the intrinsic physics interest in the measurement of these Standard Model processes, as well as their role as backgrounds for New Physics searches.

  6. Modelling of the $t\\bar{t}H$ and $t\\bar{t}V$ $(V=W,Z)$ processes for $\\sqrt{s}=13$ TeV ATLAS analyses

    CERN Document Server

    The ATLAS collaboration

    2016-01-01

    Production of top quark pairs in association with heavy Standard Model bosons is important both as a signal and a background in several ATLAS analyses. Strong constraints on such processes cannot at present be obtained from data, and therefore their modelling by Monte Carlo simulation as well as the associated uncertainties are important. This note documents the Monte Carlo samples currently being used in ATLAS for the $t\\bar{t}H$ and $t\\bar{t}V$ ($V=W,Z$ vector bosons) processes for $\\sqrt{s}=13$ TeV proton-proton collisions.

  7. Analysis and modeling of wafer-level process variability in 28 nm FD-SOI using split C-V measurements

    Science.gov (United States)

    Pradeep, Krishna; Poiroux, Thierry; Scheer, Patrick; Juge, André; Gouget, Gilles; Ghibaudo, Gérard

    2018-07-01

    This work details the analysis of wafer level global process variability in 28 nm FD-SOI using split C-V measurements. The proposed approach initially evaluates the native on wafer process variability using efficient extraction methods on split C-V measurements. The on-wafer threshold voltage (VT) variability is first studied and modeled using a simple analytical model. Then, a statistical model based on the Leti-UTSOI compact model is proposed to describe the total C-V variability in different bias conditions. This statistical model is finally used to study the contribution of each process parameter to the total C-V variability.

  8. Satellite image simulations for model-supervised, dynamic retrieval of crop type and land use intensity

    Science.gov (United States)

    Bach, H.; Klug, P.; Ruf, T.; Migdall, S.; Schlenz, F.; Hank, T.; Mauser, W.

    2015-04-01

    To support food security, information products about the actual cropping area per crop type, the current status of agricultural production and estimated yields, as well as the sustainability of the agricultural management are necessary. Based on this information, well-targeted land management decisions can be made. Remote sensing is in a unique position to contribute to this task as it is globally available and provides a plethora of information about current crop status. M4Land is a comprehensive system in which a crop growth model (PROMET) and a reflectance model (SLC) are coupled in order to provide these information products by analyzing multi-temporal satellite images. SLC uses modelled surface state parameters from PROMET, such as leaf area index or phenology of different crops to simulate spatially distributed surface reflectance spectra. This is the basis for generating artificial satellite images considering sensor specific configurations (spectral bands, solar and observation geometries). Ensembles of model runs are used to represent different crop types, fertilization status, soil colour and soil moisture. By multi-temporal comparisons of simulated and real satellite images, the land cover/crop type can be classified in a dynamically, model-supervised way and without in-situ training data. The method is demonstrated in an agricultural test-site in Bavaria. Its transferability is studied by analysing PROMET model results for the rest of Germany. Especially the simulated phenological development can be verified on this scale in order to understand whether PROMET is able to adequately simulate spatial, as well as temporal (intra- and inter-season) crop growth conditions, a prerequisite for the model-supervised approach. This sophisticated new technology allows monitoring of management decisions on the field-level using high resolution optical data (presently RapidEye and Landsat). The M4Land analysis system is designed to integrate multi-mission data and is

  9. HSP v2: Haptic Signal Processing with Extensions for Physical Modeling

    DEFF Research Database (Denmark)

    Overholt, Daniel; Kontogeorgakopoulos, Alexandros; Berdahl, Edgar

    2010-01-01

    The Haptic Signal Processing (HSP) platform aims to enable musicians to easily design and perform with digital haptic musical instruments [1]. In this paper, we present some new objects introduced in version v2 for modeling of musical dynamical systems such as resonators and vibrating strings. To....... To our knowledge, this is the first time that these diverse physical modeling elements have all been made available for a modular, real-time haptics platform....

  10. Modeling of Processing-Induced Pore Morphology in an Additively-Manufactured Ti-6Al-4V Alloy.

    Science.gov (United States)

    Kabir, Mohammad Rizviul; Richter, Henning

    2017-02-08

    A selective laser melting (SLM)-based, additively-manufactured Ti-6Al-4V alloy is prone to the accumulation of undesirable defects during layer-by-layer material build-up. Defects in the form of complex-shaped pores are one of the critical issues that need to be considered during the processing of this alloy. Depending on the process parameters, pores with concave or convex boundaries may occur. To exploit the full potential of additively-manufactured Ti-6Al-4V, the interdependency between the process parameters, pore morphology, and resultant mechanical properties, needs to be understood. By incorporating morphological details into numerical models for micromechanical analyses, an in-depth understanding of how these pores interact with the Ti-6Al-4V microstructure can be gained. However, available models for pore analysis lack a realistic description of both the Ti-6Al-4V grain microstructure, and the pore geometry. To overcome this, we propose a comprehensive approach for modeling and discretizing pores with complex geometry, situated in a polycrystalline microstructure. In this approach, the polycrystalline microstructure is modeled by means of Voronoi tessellations, and the complex pore geometry is approximated by strategically combining overlapping spheres of varied sizes. The proposed approach provides an elegant way to model the microstructure of SLM-processed Ti-6Al-4V containing pores or crack-like voids, and makes it possible to investigate the relationship between process parameters, pore morphology, and resultant mechanical properties in a finite-element-based simulation framework.

  11. Modeling of Processing-Induced Pore Morphology in an Additively-Manufactured Ti-6Al-4V Alloy

    Directory of Open Access Journals (Sweden)

    Mohammad Rizviul Kabir

    2017-02-01

    Full Text Available A selective laser melting (SLM-based, additively-manufactured Ti-6Al-4V alloy is prone to the accumulation of undesirable defects during layer-by-layer material build-up. Defects in the form of complex-shaped pores are one of the critical issues that need to be considered during the processing of this alloy. Depending on the process parameters, pores with concave or convex boundaries may occur. To exploit the full potential of additively-manufactured Ti-6Al-4V, the interdependency between the process parameters, pore morphology, and resultant mechanical properties, needs to be understood. By incorporating morphological details into numerical models for micromechanical analyses, an in-depth understanding of how these pores interact with the Ti-6Al-4V microstructure can be gained. However, available models for pore analysis lack a realistic description of both the Ti-6Al-4V grain microstructure, and the pore geometry. To overcome this, we propose a comprehensive approach for modeling and discretizing pores with complex geometry, situated in a polycrystalline microstructure. In this approach, the polycrystalline microstructure is modeled by means of Voronoi tessellations, and the complex pore geometry is approximated by strategically combining overlapping spheres of varied sizes. The proposed approach provides an elegant way to model the microstructure of SLM-processed Ti-6Al-4V containing pores or crack-like voids, and makes it possible to investigate the relationship between process parameters, pore morphology, and resultant mechanical properties in a finite-element-based simulation framework.

  12. METODOLOGIJA ZA DIMENZIONIRANJE NESEMAFORIZIRANIH NIVOJSKIH KRIŽIŠČ

    OpenAIRE

    Vogrinec, Petra

    2010-01-01

    Promet je pomemben dejavnik v razvoju človeštva. Omogoča transport ljudi, tovora in informacij. Ker se cestni promet iz leta v leto povečuje, prihaja do večjih zamud in kolon na cestah in v križiščih. Če so povprečne zamude prevelike, je treba izvesti ukrepe za povečanje prepustnosti ceste oziroma križišča. V diplomi je predstavljena metodologija za dimenzioniranje nesemaforiziranih nivojskih križišč. Poudarek je na križišču glavne in stranske ceste. Izračun je narejen po »Highway Capacity...

  13. On the Use of Variability Operations in the V-Modell XT Software Process Line

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Méndez Fernández, Daniel; Ternité, Thomas

    2016-01-01

    . In this article, we present a study on the feasibility of variability operations to support the development of software process lines in the context of the V-Modell XT. We analyze which variability operations are defined and practically used. We provide an initial catalog of variability operations...... as an improvement proposal for other process models. Our findings show that 69 variability operation types are defined across several metamodel versions of which, however, 25 remain unused. The found variability operations allow for systematically modifying the content of process model elements and the process......Software process lines provide a systematic approach to develop and manage software processes. It defines a reference process containing general process assets, whereas a well-defined customization approach allows process engineers to create new process variants, e.g., by extending or modifying...

  14. Monte Carlo modeling of Standard Model multi-boson production processes for $\\sqrt{s} = 13$ TeV ATLAS analyses

    CERN Document Server

    Li, Shu; The ATLAS collaboration

    2017-01-01

    Proceeding for the poster presentation at LHCP2017, Shanghai, China on the topic of "Monte Carlo modeling of Standard Model multi-boson production processes for $\\sqrt{s} = 13$ TeV ATLAS analyses" (ATL-PHYS-SLIDE-2017-265 https://cds.cern.ch/record/2265389) Deadline: 01/09/2017

  15. Monte Carlo modeling of Standard Model multi-boson production processes for √s = 13 TeV ATLAS analyses

    CERN Document Server

    Li, Shu; The ATLAS collaboration

    2017-01-01

    We present the Monte Carlo(MC) setup used by ATLAS to model multi-boson processes in √s = 13 TeV proton-proton collisions. The baseline Monte Carlo generators are compared with each other in key kinematic distributions of the processes under study. Sample normalization and systematic uncertainties are discussed.

  16. CFD modeling of combustion processes using KIVA3V Code with partially stirred reactor model for turbulence-combustion interactions

    International Nuclear Information System (INIS)

    Jarnicki, R.; Sobiesiak, A.

    2002-01-01

    In order to solve the averaged conservation equations for turbulent reacting flow one is faced with a task of specifying the averaged chemical reaction rate. This is due to turbulence influence on the mean reaction rates that appear in the species concentration Reynolds-averaged equation. In order to investigate the Partially Stirred Reactor (PaSR) combustion model capabilities, a CFD modeling using KIVA3V Code with the PaSR model of two very different combustion processes, was performed. Experimental results were compared with modeling

  17. Standard Model processes

    CERN Document Server

    Mangano, M.L.; Aguilar-Saavedra, Juan Antonio; Alekhin, S.; Badger, S.; Bauer, C.W.; Becher, T.; Bertone, V.; Bonvini, M.; Boselli, S.; Bothmann, E.; Boughezal, R.; Cacciari, M.; Carloni Calame, C.M.; Caola, F.; Campbell, J.M.; Carrazza, S.; Chiesa, M.; Cieri, L.; Cimaglia, F.; Febres Cordero, F.; Ferrarese, P.; D'Enterria, D.; Ferrera, G.; Garcia i Tormo, X.; Garzelli, M.V.; Germann, E.; Hirschi, V.; Han, T.; Ita, H.; Jäger, B.; Kallweit, S.; Karlberg, A.; Kuttimalai, S.; Krauss, F.; Larkoski, A.J.; Lindert, J.; Luisoni, G.; Maierhöfer, P.; Mattelaer, O.; Martinez, H.; Moch, S.; Montagna, G.; Moretti, M.; Nason, P.; Nicrosini, O.; Oleari, C.; Pagani, D.; Papaefstathiou, A.; Petriello, F.; Piccinini, F.; Pierini, M.; Pierog, T.; Pozzorini, S.; Re, E.; Robens, T.; Rojo, J.; Ruiz, R.; Sakurai, K.; Salam, G.P.; Salfelder, L.; Schönherr, M.; Schulze, M.; Schumann, S.; Selvaggi, M.; Shivaji, A.; Siodmok, A.; Skands, P.; Torrielli, P.; Tramontano, F.; Tsinikos, I.; Tweedie, B.; Vicini, A.; Westhoff, S.; Zaro, M.; Zeppenfeld, D.; CERN. Geneva. ATS Department

    2017-06-22

    This report summarises the properties of Standard Model processes at the 100 TeV pp collider. We document the production rates and typical distributions for a number of benchmark Standard Model processes, and discuss new dynamical phenomena arising at the highest energies available at this collider. We discuss the intrinsic physics interest in the measurement of these Standard Model processes, as well as their role as backgrounds for New Physics searches.

  18. Terrestrial ecosystem process model Biome-BGCMuSo v4.0: summary of improvements and new modeling possibilities

    Science.gov (United States)

    Hidy, Dóra; Barcza, Zoltán; Marjanović, Hrvoje; Zorana Ostrogović Sever, Maša; Dobor, Laura; Gelybó, Györgyi; Fodor, Nándor; Pintér, Krisztina; Churkina, Galina; Running, Steven; Thornton, Peter; Bellocchi, Gianni; Haszpra, László; Horváth, Ferenc; Suyker, Andrew; Nagy, Zoltán

    2016-12-01

    The process-based biogeochemical model Biome-BGC was enhanced to improve its ability to simulate carbon, nitrogen, and water cycles of various terrestrial ecosystems under contrasting management activities. Biome-BGC version 4.1.1 was used as a base model. Improvements included addition of new modules such as the multilayer soil module, implementation of processes related to soil moisture and nitrogen balance, soil-moisture-related plant senescence, and phenological development. Vegetation management modules with annually varying options were also implemented to simulate management practices of grasslands (mowing, grazing), croplands (ploughing, fertilizer application, planting, harvesting), and forests (thinning). New carbon and nitrogen pools have been defined to simulate yield and soft stem development of herbaceous ecosystems. The model version containing all developments is referred to as Biome-BGCMuSo (Biome-BGC with multilayer soil module; in this paper, Biome-BGCMuSo v4.0 is documented). Case studies on a managed forest, cropland, and grassland are presented to demonstrate the effect of model developments on the simulation of plant growth as well as on carbon and water balance.

  19. A design for a V and V and UQ discovery process

    International Nuclear Information System (INIS)

    Knupp, Patrick Michael; Urbina, Angel

    2011-01-01

    There is currently sparse literature on how to implement systematic and comprehensive processes for modern V and V/UQ (VU) within large computational simulation projects. Important design requirements have been identified in order to construct a viable 'system' of processes. Significant processes that are needed include discovery, accumulation, and assessment. A preliminary design is presented for a VU Discovery process that accounts for an important subset of the requirements. The design uses a hierarchical approach to set context and a series of place-holders that identify the evidence and artifacts that need to be created in order to tell the VU story and to perform assessments. The hierarchy incorporates VU elements from a Predictive Capability Maturity Model and uses questionnaires to define critical issues in VU. The place-holders organize VU data within a central repository that serves as the official VU record of the project. A review process ensures that those who will contribute to the record have agreed to provide the evidence identified by the Discovery process. VU expertise is an essential part of this process and ensures that the roadmap provided by the Discovery process is adequate. Both the requirements and the design were developed to support the Nuclear Energy Advanced Modeling and Simulation Waste project, which is developing a set of advanced codes for simulating the performance of nuclear waste storage sites. The Waste project served as an example to keep the design of the VU Discovery process grounded in practicalities. However, the system is represented abstractly so that it can be applied to other M and S projects.

  20. An ecological vegetation-activated sludge process (V-ASP) for decentralized wastewater treatment: system development, treatment performance, and mathematical modeling.

    Science.gov (United States)

    Yuan, Jiajia; Dong, Wenyi; Sun, Feiyun; Li, Pu; Zhao, Ke

    2016-05-01

    An environment-friendly decentralized wastewater treatment process that is comprised of activated sludge process (ASP) and wetland vegetation, named as vegetation-activated sludge process (V-ASP), was developed for decentralized wastewater treatment. The long-term experimental results evidenced that the vegetation sequencing batch reactor (V-SBR) process had consistently stable higher removal efficiencies of organic substances and nutrients from domestic wastewater compared with traditional sequencing batch reactor (SBR). The vegetation allocated into V-SBR system could not only remove nutrients through its vegetation transpiration ratio but also provide great surface area for microorganism activity enhancement. This high vegetation transpiration ratio enhanced nutrients removal effectiveness from wastewater mainly by flux enhancement, oxygen and substrate transportation acceleration, and vegetation respiration stimulation. A mathematical model based on ASM2d was successfully established by involving the specific function of vegetation to simulate system performance. The simulation results on the influence of operational parameters on V-ASP treatment effectiveness demonstrated that V-SBR had a high resistance to seasonal temperature fluctuations and influent loading shocking.

  1. Parameters Determination of Yoshida Uemori Model Through Optimization Process of Cyclic Tension-Compression Test and V-Bending Springback

    Directory of Open Access Journals (Sweden)

    Serkan Toros

    Full Text Available Abstract In recent years, the studies on the enhancement of the prediction capability of the sheet metal forming simulations have increased remarkably. Among the used models in the finite element simulations, the yield criteria and hardening models have a great importance for the prediction of the formability and springback. The required model parameters are determined by using the several test results, i.e. tensile, compression, biaxial stretching tests (bulge test and cyclic tests (tension-compression. In this study, the Yoshida-Uemori (combined isotropic and kinematic hardening model is used to determine the performance of the springback prediction. The model parameters are determined by the optimization processes of the cyclic test by finite element simulations. However, in the study besides the cyclic tests, the model parameters are also evaluated by the optimization process of both cyclic and V-die bending simulations. The springback angle predictions with the model parameters obtained by the optimization of both cyclic and V-die bending simulations are found to mimic the experimental results in a better way than those obtained from only cyclic tests. However, the cyclic simulation results are found to be close enough to the experimental results.

  2. Software reliability evaluation of digital plant protection system development process using V and V

    International Nuclear Information System (INIS)

    Lee, Na Young; Hwang, Il Soon; Seong, Seung Hwan; Oh, Seung Rok

    2001-01-01

    In the nuclear power industry, digital technology has been introduced recently for the Instrumentation and Control (I and C) of reactor systems. For its application to the safety critical system such as Reactor Protection System(RPS), a reliability assessment is indispensable. Unlike traditional reliability models, software reliability is hard to evaluate, and should be evaluated throughout development lifecycle. In the development process of Digital Plant Protection System(DPPS), the concept of verification and validation (V and V) was introduced to assure the quality of the product. Also, test should be performed to assure the reliability. Verification procedure with model checking is relatively well defined, however, test is labor intensive and not well organized. In this paper, we developed the methodological process of combining the verification with validation test case generation. For this, we used PVS for the table specification and for the theorem proving. As a result, we could not only save time to design test case but also get more effective and complete verification related test case set. Add to this, we could extract some meaningful factors useful for the reliability evaluation both from the V and V and verification combined tests

  3. Tudengite hea esinemine Willem C. Vis'i nimelisel harjutuskohtuvõistlusel

    Index Scriptorium Estoniae

    2013-01-01

    XX Willem C. Vis’i nimelise rahvusvahelise kaubandusõiguse harjutuskohtuvõistlusel Viinis saavutas Tartu Ülikooli õigusteaduskonna võistkond (Kristina Promet, Jaana Lints, Edgar-Kaj Velbri ja Sandra-Kristin Noot) 290 ülikooli seast 84. koha

  4. V and V-based remaining fault estimation model for safety–critical software of a nuclear power plant

    International Nuclear Information System (INIS)

    Eom, Heung-seop; Park, Gee-yong; Jang, Seung-cheol; Son, Han Seong; Kang, Hyun Gook

    2013-01-01

    Highlights: ► A software fault estimation model based on Bayesian Nets and V and V. ► Use of quantified data derived from qualitative V and V results. ► Faults insertion and elimination process was modeled in the context of probability. ► Systematically estimates the expected number of remaining faults. -- Abstract: Quantitative software reliability measurement approaches have some limitations in demonstrating the proper level of reliability in cases of safety–critical software. One of the more promising alternatives is the use of software development quality information. Particularly in the nuclear industry, regulatory bodies in most countries use both probabilistic and deterministic measures for ensuring the reliability of safety-grade digital computers in NPPs. The point of deterministic criteria is to assess the whole development process and its related activities during the software development life cycle for the acceptance of safety–critical software. In addition software Verification and Validation (V and V) play an important role in this process. In this light, we propose a V and V-based fault estimation method using Bayesian Nets to estimate the remaining faults for safety–critical software after the software development life cycle is completed. By modeling the fault insertion and elimination processes during the whole development phases, the proposed method systematically estimates the expected number of remaining faults.

  5. ORCHIDEE-CROP (v0), a new process-based agro-land surface model: model description and evaluation over Europe

    Science.gov (United States)

    Wu, X.; Vuichard, N.; Ciais, P.; Viovy, N.; de Noblet-Ducoudré, N.; Wang, X.; Magliulo, V.; Wattenbach, M.; Vitale, L.; Di Tommasi, P.; Moors, E. J.; Jans, W.; Elbers, J.; Ceschia, E.; Tallec, T.; Bernhofer, C.; Grünwald, T.; Moureaux, C.; Manise, T.; Ligne, A.; Cellier, P.; Loubet, B.; Larmanou, E.; Ripoche, D.

    2016-03-01

    The response of crops to changing climate and atmospheric CO2 concentration ([CO2]) could have large effects on food production, and impact carbon, water, and energy fluxes, causing feedbacks to the climate. To simulate the response of temperate crops to changing climate and [CO2], which accounts for the specific phenology of crops mediated by management practice, we describe here the development of a process-oriented terrestrial biogeochemical model named ORCHIDEE-CROP (v0), which integrates a generic crop phenology and harvest module, and a very simple parameterization of nitrogen fertilization, into the land surface model (LSM) ORCHIDEEv196, in order to simulate biophysical and biochemical interactions in croplands, as well as plant productivity and harvested yield. The model is applicable for a range of temperate crops, but is tested here using maize and winter wheat, with the phenological parameterizations of two European varieties originating from the STICS agronomical model. We evaluate the ORCHIDEE-CROP (v0) model against eddy covariance and biometric measurements at seven winter wheat and maize sites in Europe. The specific ecosystem variables used in the evaluation are CO2 fluxes (net ecosystem exchange, NEE), latent heat, and sensible heat fluxes. Additional measurements of leaf area index (LAI) and aboveground biomass and yield are used as well. Evaluation results revealed that ORCHIDEE-CROP (v0) reproduced the observed timing of crop development stages and the amplitude of the LAI changes. This is in contrast to ORCHIDEEv196 where, by default, crops have the same phenology as grass. A halving of the root mean square error for LAI from 2.38 ± 0.77 to 1.08 ± 0.34 m2 m-2 was obtained when ORCHIDEEv196 and ORCHIDEE-CROP (v0) were compared across the seven study sites. Improved crop phenology and carbon allocation led to a good match between modeled and observed aboveground biomass (with a normalized root mean squared error (NRMSE) of 11.0-54.2 %), crop

  6. A process improvement model for software verification and validation

    Science.gov (United States)

    Callahan, John; Sabolish, George

    1994-01-01

    We describe ongoing work at the NASA Independent Verification and Validation (IV&V) Facility to establish a process improvement model for software verification and validation (V&V) organizations. This model, similar to those used by some software development organizations, uses measurement-based techniques to identify problem areas and introduce incremental improvements. We seek to replicate this model for organizations involved in V&V on large-scale software development projects such as EOS and space station. At the IV&V Facility, a university research group and V&V contractors are working together to collect metrics across projects in order to determine the effectiveness of V&V and improve its application. Since V&V processes are intimately tied to development processes, this paper also examines the repercussions for development organizations in large-scale efforts.

  7. ORCHIDEE-CROP (v0), a new process based Agro-Land Surface Model: model description and evaluation over Europe

    Science.gov (United States)

    Wu, X.; Vuichard, N.; Ciais, P.; Viovy, N.; de Noblet-Ducoudré, N.; Wang, X.; Magliulo, V.; Wattenbach, M.; Vitale, L.; Di Tommasi, P.; Moors, E. J.; Jans, W.; Elbers, J.; Ceschia, E.; Tallec, T.; Bernhofer, C.; Grünwald, T.; Moureaux, C.; Manise, T.; Ligne, A.; Cellier, P.; Loubet, B.; Larmanou, E.; Ripoche, D.

    2015-06-01

    The responses of crop functioning to changing climate and atmospheric CO2 concentration ([CO2]) could have large effects on food production, and impact carbon, water and energy fluxes, causing feedbacks to climate. To simulate the responses of temperate crops to changing climate and [CO2], accounting for the specific phenology of crops mediated by management practice, we present here the development of a process-oriented terrestrial biogeochemical model named ORCHIDEE-CROP (v0), which integrates a generic crop phenology and harvest module and a very simple parameterization of nitrogen fertilization, into the land surface model (LSM) ORCHIDEEv196, in order to simulate biophysical and biochemical interactions in croplands, as well as plant productivity and harvested yield. The model is applicable for a range of temperate crops, but it is tested here for maize and winter wheat, with the phenological parameterizations of two European varieties originating from the STICS agronomical model. We evaluate the ORCHIDEE-CROP (v0) model against eddy covariance and biometric measurements at 7 winter wheat and maize sites in Europe. The specific ecosystem variables used in the evaluation are CO2 fluxes (NEE), latent heat and sensible heat fluxes. Additional measurements of leaf area index (LAI), aboveground biomass and yield are used as well. Evaluation results reveal that ORCHIDEE-CROP (v0) reproduces the observed timing of crop development stages and the amplitude of pertaining LAI changes in contrast to ORCHIDEEv196 in which by default crops have the same phenology than grass. A near-halving of the root mean square error of LAI from 2.38 ± 0.77 to 1.08 ± 0.34 m2 m-2 is obtained between ORCHIDEEv196 and ORCHIDEE-CROP (v0) across the 7 study sites. Improved crop phenology and carbon allocation lead to a general good match between modelled and observed aboveground biomass (with a normalized root mean squared error (NRMSE) of 11.0-54.2 %), crop yield, as well as of the daily

  8. Robot vision language RVL/V: An integration scheme of visual processing and manipulator control

    International Nuclear Information System (INIS)

    Matsushita, T.; Sato, T.; Hirai, S.

    1984-01-01

    RVL/V is a robot vision language designed to write a program for visual processing and manipulator control of a hand-eye system. This paper describes the design of RVL/V and the current implementation of the system. Visual processing is performed on one-dimensional range data of the object surface. Model-based instructions execute object detection, measurement and view control. The hierarchy of visual data and processing is introduced to give RVL/V generality. A new scheme to integrate visual information and manipulator control is proposed. The effectiveness of the model-based visual processing scheme based on profile data is demonstrated by a hand-eye experiment

  9. Temperature dependence of frequency dispersion in III–V metal-oxide-semiconductor C-V and the capture/emission process of border traps

    Energy Technology Data Exchange (ETDEWEB)

    Vais, Abhitosh, E-mail: Abhitosh.Vais@imec.be; Martens, Koen; DeMeyer, Kristin [Department of Electrical Engineering, KU Leuven, B-3000 Leuven (Belgium); IMEC, Kapeldreef 75, B-3001 Leuven (Belgium); Lin, Han-Chung; Ivanov, Tsvetan; Collaert, Nadine; Thean, Aaron [IMEC, Kapeldreef 75, B-3001 Leuven (Belgium); Dou, Chunmeng [Frontier Research Center, Tokyo Institute of Technology, Yokohama 226-8502 (Japan); Xie, Qi; Maes, Jan [ASM International, B-3001 Leuven (Belgium); Tang, Fu; Givens, Michael [ASM International, Phoenix, Arizona 85034-7200 (United States); Raskin, Jean-Pierre [Institute of Information and Communication Technologies, Electronics and Applied Mathematics, Universiteé Catholique de Louvain, B-1348 Louvain-la-Neuve (Belgium)

    2015-08-03

    This paper presents a detailed investigation of the temperature dependence of frequency dispersion observed in capacitance-voltage (C-V) measurements of III-V metal-oxide-semiconductor (MOS) devices. The dispersion in the accumulation region of the capacitance data is found to change from 4%–9% (per decade frequency) to ∼0% when the temperature is reduced from 300 K to 4 K in a wide range of MOS capacitors with different gate dielectrics and III-V substrates. We show that such significant temperature dependence of C-V frequency dispersion cannot be due to the temperature dependence of channel electrostatics, i.e., carrier density and surface potential. We also show that the temperature dependence of frequency dispersion, and hence, the capture/emission process of border traps can be modeled by a combination of tunneling and a “temperature-activated” process described by a non-radiative multi-phonon model, instead of a widely believed single-step elastic tunneling process.

  10. Process-oriented microstructure evolution of V{sub ss}-V{sub 3}Si-V{sub 5}SiB{sub 2}; Prozessabhaengige Mikrostrukturausbildung von V{sub ss}-V{sub 3}Si-V{sub 5}SiB{sub 2}-Werkstoffen

    Energy Technology Data Exchange (ETDEWEB)

    Krueger, Manja; Koeppe-Grabow, Birte [Magdeburg Univ. (Germany)

    2017-05-15

    Vanadium silicide alloys are potentially interesting high temperature materials, since they combine high mechanical strength at temperatures of up to 1 000 C with a low density. In this study, the microstructures of innovative V-Si-B high temperature materials are examined using different analytical methods. The selected V-9Si-13B model alloy was manufactured using a powder metallurgical process route as well as an ingot metallurgical process. The alloys show a vanadium solid solution phase as well as the high-strength silicide phases V{sub 3}Si and V{sub 5}SiB{sub 2}. Especially for the powder metallurgically fabricated alloy, showing finely dispersed phases, the quantification of microstructural constituents is difficult. The phases, however, can be separated from one another via computer tomography.

  11. Qualitative simulation in formal process modelling

    International Nuclear Information System (INIS)

    Sivertsen, Elin R.

    1999-01-01

    In relation to several different research activities at the OECD Halden Reactor Project, the usefulness of formal process models has been identified. Being represented in some appropriate representation language, the purpose of these models is to model process plants and plant automatics in a unified way to allow verification and computer aided design of control strategies. The present report discusses qualitative simulation and the tool QSIM as one approach to formal process models. In particular, the report aims at investigating how recent improvements of the tool facilitate the use of the approach in areas like process system analysis, procedure verification, and control software safety analysis. An important long term goal is to provide a basis for using qualitative reasoning in combination with other techniques to facilitate the treatment of embedded programmable systems in Probabilistic Safety Analysis (PSA). This is motivated from the potential of such a combination in safety analysis based on models comprising both software, hardware, and operator. It is anticipated that the research results from this activity will benefit V and V in a wide variety of applications where formal process models can be utilized. Examples are operator procedures, intelligent decision support systems, and common model repositories (author) (ml)

  12. V&V framework

    Energy Technology Data Exchange (ETDEWEB)

    Hills, Richard G. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Maniaci, David Charles [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Naughton, Jonathan W. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-09-01

    A Verification and Validation (V&V) framework is presented for the development and execution of coordinated modeling and experimental program s to assess the predictive capability of computational models of complex systems through focused, well structured, and formal processes.The elements of the framework are based on established V&V methodology developed by various organizations including the Department of Energy, National Aeronautics and Space Administration, the American Institute of Aeronautics and Astronautics, and the American Society of Mechanical Engineers. Four main topics are addressed: 1) Program planning based on expert elicitation of the modeling physics requirements, 2) experimental design for model assessment, 3) uncertainty quantification for experimental observations and computational model simulations, and 4) assessment of the model predictive capability. The audience for this document includes program planners, modelers, experimentalist, V &V specialist, and customers of the modeling results.

  13. Recurrent connectivity can account for the dynamics of disparity processing in V1

    Science.gov (United States)

    Samonds, Jason M.; Potetz, Brian R.; Tyler, Christopher W.; Lee, Tai Sing

    2013-01-01

    Disparity tuning measured in the primary visual cortex (V1) is described well by the disparity energy model, but not all aspects of disparity tuning are fully explained by the model. Such deviations from the disparity energy model provide us with insight into how network interactions may play a role in disparity processing and help to solve the stereo correspondence problem. Here, we propose a neuronal circuit model with recurrent connections that provides a simple account of the observed deviations. The model is based on recurrent connections inferred from neurophysiological observations on spike timing correlations, and is in good accord with existing data on disparity tuning dynamics. We further performed two additional experiments to test predictions of the model. First, we increased the size of stimuli to drive more neurons and provide a stronger recurrent input. Our model predicted sharper disparity tuning for larger stimuli. Second, we displayed anti-correlated stereograms, where dots of opposite luminance polarity are matched between the left- and right-eye images and result in inverted disparity tuning in the disparity energy model. In this case, our model predicted reduced sharpening and strength of inverted disparity tuning. For both experiments, the dynamics of disparity tuning observed from the neurophysiological recordings in macaque V1 matched model simulation predictions. Overall, the results of this study support the notion that, while the disparity energy model provides a primary account of disparity tuning in V1 neurons, neural disparity processing in V1 neurons is refined by recurrent interactions among elements in the neural circuit. PMID:23407952

  14. Diseño de multiplexores hibridos para sistemas de comunicaciones ópticas mediante la concatenación de dispositivos selectivos en longitud de onda

    OpenAIRE

    Rodríguez Horche, Paloma; López-Amo Sainz, Manuel; Muriel Fernández, Miguel Ángel; Martín Pereda, José Antonio

    1987-01-01

    La creciente demandada de información por parte del abonado hace necesaria la inclusión de nuevos servicios como la transmisión da datos de alta y baja velocidad, señales de vídeo, etc. La única solución que promete viabilidad es la que puede desarrollarse con Fibra Óptica.

  15. Segmentation of kidney using C-V model and anatomy priors

    Science.gov (United States)

    Lu, Jinghua; Chen, Jie; Zhang, Juan; Yang, Wenjia

    2007-12-01

    This paper presents an approach for kidney segmentation on abdominal CT images as the first step of a virtual reality surgery system. Segmentation for medical images is often challenging because of the objects' complicated anatomical structures, various gray levels, and unclear edges. A coarse to fine approach has been applied in the kidney segmentation using Chan-Vese model (C-V model) and anatomy prior knowledge. In pre-processing stage, the candidate kidney regions are located. Then C-V model formulated by level set method is applied in these smaller ROI, which can reduce the calculation complexity to a certain extent. At last, after some mathematical morphology procedures, the specified kidney structures have been extracted interactively with prior knowledge. The satisfying results on abdominal CT series show that the proposed approach keeps all the advantages of C-V model and overcome its disadvantages.

  16. The V sign in lateral talar process fractures: an experimental study using a foot and ankle model.

    Science.gov (United States)

    Jentzsch, Thorsten; Hasler, Anita; Renner, Niklas; Peterhans, Manuel; Sutter, Reto; Espinosa, Norman; Wirth, Stephan H

    2017-07-03

    Lateral talar process fractures (LTPF) are often missed on conventional radiographs. A positive V sign is an interruption of the contour of the LTP. It has been suggested, but not proven to be pathognomonic for LTPF. The objective was to study whether the V sign is pathognomonic for LTPF and if it can be properly assessed in different ankle positions and varying fracture types. An experimental study was conducted. Two investigators assessed lateral radiographs (n = 108) of a foot and ankle model. The exposure variables were different ankle positions and fracture types. The primary outcome was the correct detection of a V sign. The secondary outcomes were the detection of the V sign depending on ankle position and fracture type as well as the uncertainty. The interobserver agreement on the V sign and type of fracture were fair (κ = 0.35, 95% CI 0.18-0.53, p ankle position and fracture type showed significant better results with increasing inversion (p = 0.035 and p = 0.011) and type B fractures (p = 0.001 and p = 0.013). The V sign may not be pathognomonic and is not recommended as the only modality for the detection of LTPF. It is better visualized with inversion, but does not depend on plantar flexion or internal rotation. It is also better seen in type B fractures. It is difficult to detect and investigator-dependent. It may be helpful in a clinical setting to point into a direction, but a CT scan may be used if in doubt about a LTPF.

  17. V and V Efforts of Auroral Precipitation Models: Preliminary Results

    Science.gov (United States)

    Zheng, Yihua; Kuznetsova, Masha; Rastaetter, Lutz; Hesse, Michael

    2011-01-01

    Auroral precipitation models have been valuable both in terms of space weather applications and space science research. Yet very limited testing has been performed regarding model performance. A variety of auroral models are available, including empirical models that are parameterized by geomagnetic indices or upstream solar wind conditions, now casting models that are based on satellite observations, or those derived from physics-based, coupled global models. In this presentation, we will show our preliminary results regarding V&V efforts of some of the models.

  18. ENDF/B-V processing programs

    International Nuclear Information System (INIS)

    DayDay, N.

    1980-07-01

    A description and operating instructions are supplied for the following ENDF/B-V Processing Programs: CHECKER, CRECT, STNDRD, FIZCON, PSYCHE, RESEND, INTER, INTEND, SUMRIZ, PLOTEF, LSTFCV, RIGEL. These programs can be obtained on magnetic tape, free of charge, from the IAEA Nuclear Data Section. (author)

  19. DEJAVNIKI ZDRAVSTVENEGA TVEGANJA NA DELOVNEM MESTU PATRONAŽNE MEDICINSKE SESTRE

    OpenAIRE

    Vratarič, Marija Sabrina

    2015-01-01

    Patronažno varstvo zajema izredno pestro simptomatiko in problematiko. Potrebe po patronažni zdravstveni negi se po napovedih še povečujejo. Trendi sodobne družbe usmerjajo oskrbo posameznika tako, da ostane v domačem okolju čim dlje aktiven, zdrav in sposoben samooskrbe. Več vrst nasilja, promet in vremenske razmere, pravočasnost informacij ter neprimerno delovno okolje so dejavniki tveganja, ki smo jih v naši magistrski nalogi izpostavili in se v raziskavi tudi izkažejo za obremenjujoč...

  20. Analiza izpustov dizelskih motorjev

    OpenAIRE

    MURN, GREGA

    2016-01-01

    V svojem diplomskem delu z naslovom »Analiza izpustov dizelskih motorjev«, sem bralcu skušal približati današnjo problematiko v zvezi z onesnaževanjem okolja, predvsem zaradi dizelskih izpuhov, ki jih prinaša industrija v vseh oblikah. Največji onesnaževalec okolja je absolutno promet, saj izpuhi avtomobilov vsebujejo obsežen skupek škodljivih snovi (ogljikovodiki HC, ogljikov monoksid CO, dušikovi oksidi NOx, žveplovi oksidi, trdni delci, itd.). Namen Evropskih in Ameriških standardov je ...

  1. IV&V Project Assessment Process Validation

    Science.gov (United States)

    Driskell, Stephen

    2012-01-01

    The Space Launch System (SLS) will launch NASA's Multi-Purpose Crew Vehicle (MPCV). This launch vehicle will provide American launch capability for human exploration and travelling beyond Earth orbit. SLS is designed to be flexible for crew or cargo missions. The first test flight is scheduled for December 2017. The SLS SRR/SDR provided insight into the project development life cycle. NASA IV&V ran the standard Risk Based Assessment and Portfolio Based Risk Assessment to identify analysis tasking for the SLS program. This presentation examines the SLS System Requirements Review/System Definition Review (SRR/SDR), IV&V findings for IV&V process validation correlation to/from the selected IV&V tasking and capabilities. It also provides a reusable IEEE 1012 scorecard for programmatic completeness across the software development life cycle.

  2. Model-based verification and validation of the SMAP uplink processes

    Science.gov (United States)

    Khan, M. O.; Dubos, G. F.; Tirona, J.; Standley, S.

    Model-Based Systems Engineering (MBSE) is being used increasingly within the spacecraft design community because of its benefits when compared to document-based approaches. As the complexity of projects expands dramatically with continually increasing computational power and technology infusion, the time and effort needed for verification and validation (V& V) increases geometrically. Using simulation to perform design validation with system-level models earlier in the life cycle stands to bridge the gap between design of the system (based on system-level requirements) and verifying those requirements/validating the system as a whole. This case study stands as an example of how a project can validate a system-level design earlier in the project life cycle than traditional V& V processes by using simulation on a system model. Specifically, this paper describes how simulation was added to a system model of the Soil Moisture Active-Passive (SMAP) mission's uplink process. Also discussed are the advantages and disadvantages of the methods employed and the lessons learned; which are intended to benefit future model-based and simulation-based development efforts.

  3. A process-based model for cattle manure compost windrows: Model performance and application

    Science.gov (United States)

    A model was developed and incorporated in the Integrated Farm System Model (IFSM, v.4.3) that simulates important processes occurring during windrow composting of manure. The model, documented in an accompanying paper, predicts changes in windrow properties and conditions and the resulting emissions...

  4. Working memory and decision processes in visual area V4

    Directory of Open Access Journals (Sweden)

    Benjamin eHayden

    2013-02-01

    Full Text Available Recognizing and responding to a remembered stimulus requires the coordination of perception, working memory and decision-making. To investigate the role of visual cortex in these processes, we recorded responses of single V4 neurons during performance of a delayed match-to-sample task that incorporates rapid serial visual presentation of natural images. We found that neuronal activity during the delay period after the cue but before the images depends on the identity of the remembered image and that this change persists while distractors appear. This persistent response modulation has been identified as a diagnostic criterion for putative working memory signals; our data thus suggest that working memory may involve reactivation of sensory neurons. When the remembered image reappears in the neuron’s receptive field, visually evoked responses are enhanced; this match enhancement is a diagnostic criterion for decision. One model that predicts these data is the matched filter hypothesis, which holds that during search V4 neurons change their tuning so as to match the remembered cue, and thus become detectors for that image. More generally, these results suggest that V4 neurons participate in the perceptual, working memory and decision processes that are needed to perform memory-guided decision-making.

  5. Working memory and decision processes in visual area v4.

    Science.gov (United States)

    Hayden, Benjamin Y; Gallant, Jack L

    2013-01-01

    Recognizing and responding to a remembered stimulus requires the coordination of perception, working memory, and decision-making. To investigate the role of visual cortex in these processes, we recorded responses of single V4 neurons during performance of a delayed match-to-sample task that incorporates rapid serial visual presentation of natural images. We found that neuronal activity during the delay period after the cue but before the images depends on the identity of the remembered image and that this change persists while distractors appear. This persistent response modulation has been identified as a diagnostic criterion for putative working memory signals; our data thus suggest that working memory may involve reactivation of sensory neurons. When the remembered image reappears in the neuron's receptive field, visually evoked responses are enhanced; this match enhancement is a diagnostic criterion for decision. One model that predicts these data is the matched filter hypothesis, which holds that during search V4 neurons change their tuning so as to match the remembered cue, and thus become detectors for that image. More generally, these results suggest that V4 neurons participate in the perceptual, working memory, and decision processes that are needed to perform memory-guided decision-making.

  6. Modelling floods in the Ammer catchment: limitations and challenges with a coupled meteo-hydrological model approach

    Directory of Open Access Journals (Sweden)

    R. Ludwig

    2003-01-01

    Full Text Available Numerous applications of hydrological models have shown their capability to simulate hydrological processes with a reasonable degree of certainty. For flood modelling, the quality of precipitation data — the key input parameter — is very important but often remains questionable. This paper presents a critical review of experience in the EU-funded RAPHAEL project. Different meteorological data sources were evaluated to assess their applicability for flood modelling and forecasting in the Bavarian pre-alpine catchment of the Ammer river (709 km2, for which the hydrological aspects of runoff production are described as well as the complex nature of floods. Apart from conventional rain gauge data, forecasts from several Numerical Weather Prediction Models (NWP as well as rain radar data are examined, scaled and applied within the framework of a GIS-structured and physically based hydrological model. Multi-scenario results are compared and analysed. The synergetic approach leads to promising results under certain meteorological conditions but emphasises various drawbacks. At present, NWPs are the only source of rainfall forecasts (up to 96 hours with large spatial coverage and high temporal resolution. On the other hand, the coarse spatial resolution of NWP grids cannot yet address, adequately, the heterogeneous structures of orographic rainfields in complex convective situations; hence, a major downscaling problem for mountain catchment applications is introduced. As shown for two selected Ammer flood events, a high variability in prediction accuracy has still to be accepted at present. Sensitivity analysis of both meteo-data input and hydrological model performance in terms of process description are discussed and positive conclusions have been drawn for future applications of an advanced meteo-hydro model synergy. Keywords: RAPHAEL, modelling, forecasting, model coupling, PROMET-D, TOPMODEL

  7. Nucleon-nucleon optical model for energies to 3 GeV

    International Nuclear Information System (INIS)

    Funk, A.; Von Geramb, H.V.; University of Melbourne, VIC; Amos, K.A.

    2001-01-01

    Several nucleon-nucleon potentials, Paris, Nijmegen, Argonne, and those derived by quantum inversion, which describe the NN interaction for T Lab ≤ 300 MeV are extended in their range of application as NN optical models. Extensions are made in r-space using complex separable potentials definable with a wide range of form factor options including those of boundary condition models. We use the latest phase shift analyses SP00 (FA00, WI00) of Arndt et al. from 300 MeV to 3 GeV to determine these extensions. The imaginary parts of the optical model interactions account for loss of flux into direct or resonant production processes. The optical potential approach is of particular value as it permits one to visualize fusion, and subsequent fission, of nucleons when T Lab > 2 GeV. We do so by calculating the scattering wave functions to specify the energy and radial dependences of flux losses and of probability distributions. Furthermore, half-off the energy shell t-matrices are presented as they are readily deduced with this approach. Such t-matrices are required for studies of few- and many-body nuclear reactions

  8. Building v/s Exploring Models: Comparing Learning of Evolutionary Processes through Agent-based Modeling

    Science.gov (United States)

    Wagh, Aditi

    Two strands of work motivate the three studies in this dissertation. Evolutionary change can be viewed as a computational complex system in which a small set of rules operating at the individual level result in different population level outcomes under different conditions. Extensive research has documented students' difficulties with learning about evolutionary change (Rosengren et al., 2012), particularly in terms of levels slippage (Wilensky & Resnick, 1999). Second, though building and using computational models is becoming increasingly common in K-12 science education, we know little about how these two modalities compare. This dissertation adopts agent-based modeling as a representational system to compare these modalities in the conceptual context of micro-evolutionary processes. Drawing on interviews, Study 1 examines middle-school students' productive ways of reasoning about micro-evolutionary processes to find that the specific framing of traits plays a key role in whether slippage explanations are cued. Study 2, which was conducted in 2 schools with about 150 students, forms the crux of the dissertation. It compares learning processes and outcomes when students build their own models or explore a pre-built model. Analysis of Camtasia videos of student pairs reveals that builders' and explorers' ways of accessing rules, and sense-making of observed trends are of a different character. Builders notice rules through available blocks-based primitives, often bypassing their enactment while explorers attend to rules primarily through the enactment. Moreover, builders' sense-making of observed trends is more rule-driven while explorers' is more enactment-driven. Pre and posttests reveal that builders manifest a greater facility with accessing rules, providing explanations manifesting targeted assembly. Explorers use rules to construct explanations manifesting non-targeted assembly. Interviews reveal varying degrees of shifts away from slippage in both

  9. Modelling of innovative SANEX process mal-operations

    International Nuclear Information System (INIS)

    McLachlan, F.; Taylor, R.; Whittaker, D.; Woodhead, D.; Geist, A.

    2016-01-01

    The innovative (i-) SANEX process for the separation of minor actinides from PUREX highly active raffinate is expected to employ a solvent phase comprising 0.2 M TODGA with 5 v/v% 1-octanol in an inert diluent. An initial extract / scrub section would be used to extract trivalent actinides and lanthanides from the feed whilst leaving other fission products in the aqueous phase, before the loaded solvent is contacted with a low acidity aqueous phase containing a sulphonated bis-triazinyl pyridine ligand (BTP) to effect a selective strip of the actinides, so yielding separate actinide (An) and lanthanide (Ln) product streams. This process has been demonstrated in lab scale trials at Juelich (FZJ). The SACSESS (Safety of Actinide Separation processes) project is focused on the evaluation and improvement of the safety of such future systems. A key element of this is the development of an understanding of the response of a process to upsets (mal-operations). It is only practical to study a small subset of possible mal-operations experimentally and consideration of the majority of mal-operations entails the use of a validated dynamic model of the process. Distribution algorithms for HNO_3, Am, Cm and the lanthanides have been developed and incorporated into a dynamic flowsheet model that has, so far, been configured to correspond to the extract-scrub section of the i-SANEX flowsheet trial undertaken at FZJ in 2013. Comparison is made between the steady state model results and experimental results. Results from modelling of low acidity and high temperature mal-operations are presented. (authors)

  10. Integrate-and-fire vs Poisson models of LGN input to V1 cortex: noisier inputs reduce orientation selectivity.

    Science.gov (United States)

    Lin, I-Chun; Xing, Dajun; Shapley, Robert

    2012-12-01

    One of the reasons the visual cortex has attracted the interest of computational neuroscience is that it has well-defined inputs. The lateral geniculate nucleus (LGN) of the thalamus is the source of visual signals to the primary visual cortex (V1). Most large-scale cortical network models approximate the spike trains of LGN neurons as simple Poisson point processes. However, many studies have shown that neurons in the early visual pathway are capable of spiking with high temporal precision and their discharges are not Poisson-like. To gain an understanding of how response variability in the LGN influences the behavior of V1, we study response properties of model V1 neurons that receive purely feedforward inputs from LGN cells modeled either as noisy leaky integrate-and-fire (NLIF) neurons or as inhomogeneous Poisson processes. We first demonstrate that the NLIF model is capable of reproducing many experimentally observed statistical properties of LGN neurons. Then we show that a V1 model in which the LGN input to a V1 neuron is modeled as a group of NLIF neurons produces higher orientation selectivity than the one with Poisson LGN input. The second result implies that statistical characteristics of LGN spike trains are important for V1's function. We conclude that physiologically motivated models of V1 need to include more realistic LGN spike trains that are less noisy than inhomogeneous Poisson processes.

  11. Respiratory system model for quasistatic pulmonary pressure-volume (P-V) curve: inflation-deflation loop analyses.

    Science.gov (United States)

    Amini, R; Narusawa, U

    2008-06-01

    A respiratory system model (RSM) is developed for the deflation process of a quasistatic pressure-volume (P-V) curve, following the model for the inflation process reported earlier. In the RSM of both the inflation and the deflation limb, a respiratory system consists of a large population of basic alveolar elements, each consisting of a piston-spring-cylinder subsystem. A normal distribution of the basic elements is derived from Boltzmann statistical model with the alveolar closing (opening) pressure as the distribution parameter for the deflation (inflation) process. An error minimization by the method of least squares applied to existing P-V loop data from two different data sources confirms that a simultaneous inflation-deflation analysis is required for an accurate determination of RSM parameters. Commonly used terms such as lower inflection point, upper inflection point, and compliance are examined based on the P-V equations, on the distribution function, as well as on the geometric and physical properties of the basic alveolar element.

  12. An analytical turn-on power loss model for 650-V GaN eHEMTs

    DEFF Research Database (Denmark)

    Shen, Yanfeng; Wang, Huai; Shen, Zhan

    2018-01-01

    This paper proposes an improved analytical turn-on power loss model for 650-V GaN eHEMTs. The static characteristics, i.e., the parasitic capacitances and transconductance, are firstly modeled. Then the turn-on process is divided into multiple stages and analyzed in detail; as results, the time-d......-domain solutions to the drain-source voltage and drain current are obtained. Finally, double-pulse tests are conducted to verify the proposed power loss model. This analytical model enables an accurate and fast switching behavior characterization and power loss prediction.......This paper proposes an improved analytical turn-on power loss model for 650-V GaN eHEMTs. The static characteristics, i.e., the parasitic capacitances and transconductance, are firstly modeled. Then the turn-on process is divided into multiple stages and analyzed in detail; as results, the time...

  13. Influence of ~7 keV sterile neutrino dark matter on the process of reionization

    DEFF Research Database (Denmark)

    Rudakovskyi, Anton; Iakubovskyi, Dmytro

    2016-01-01

    Recent reports of a weak unidentified emission line at ~3.5 keV found in spectra of several matter-dominated objects may give a clue to resolve the long-standing problem of dark matter. One of the best physically motivated particle candidate able to produce such an extra line is sterile neutrino...... neutrino dark matter able to produce the observed line at ~3.5 keV, to the process of reionization. By incorporating dark matter power spectra for ~7 keV sterile neutrinos into extended semi-analytical `bubble' model of reionization we obtain that such sterile neutrino dark matter would produce...

  14. Encoding model of temporal processing in human visual cortex.

    Science.gov (United States)

    Stigliani, Anthony; Jeska, Brianna; Grill-Spector, Kalanit

    2017-12-19

    How is temporal information processed in human visual cortex? Visual input is relayed to V1 through segregated transient and sustained channels in the retina and lateral geniculate nucleus (LGN). However, there is intense debate as to how sustained and transient temporal channels contribute to visual processing beyond V1. The prevailing view associates transient processing predominately with motion-sensitive regions and sustained processing with ventral stream regions, while the opposing view suggests that both temporal channels contribute to neural processing beyond V1. Using fMRI, we measured cortical responses to time-varying stimuli and then implemented a two temporal channel-encoding model to evaluate the contributions of each channel. Different from the general linear model of fMRI that predicts responses directly from the stimulus, the encoding approach first models neural responses to the stimulus from which fMRI responses are derived. This encoding approach not only predicts cortical responses to time-varying stimuli from milliseconds to seconds but also, reveals differential contributions of temporal channels across visual cortex. Consistent with the prevailing view, motion-sensitive regions and adjacent lateral occipitotemporal regions are dominated by transient responses. However, ventral occipitotemporal regions are driven by both sustained and transient channels, with transient responses exceeding the sustained. These findings propose a rethinking of temporal processing in the ventral stream and suggest that transient processing may contribute to rapid extraction of the content of the visual input. Importantly, our encoding approach has vast implications, because it can be applied with fMRI to decipher neural computations in millisecond resolution in any part of the brain. Copyright © 2017 the Author(s). Published by PNAS.

  15. La science, la littérature, le rituel: le vrai dire et le plus que mentir

    Directory of Open Access Journals (Sweden)

    Panagiotis Christias

    2005-03-01

    Full Text Available Le scientifique promet la vérité sans être en mesure d'accomplir sa promesse tandis que l'écrivain promet de mentir mais par son mensonge, il fait plus que mentir. Il ne dit certainement pas la vérité, car la vérité n'est pas sa préoccupation première. Voilà comment nous pouvons résumer les deux attitudes, condamnées à un antagonisme stérile par l'attitude positiviste moderne. Compte tenu donc de l'intention fondatrice de la science, de sa promesse donc originaire et fondatrice, la science doit dire la vérité et rien que la vérité. Ce qui, de l'aveu même de cette même science, est impossible. [...] La littérature se permet de reconstruire des mondes par-delà le vrai et le faux. [...] Un récit littéraire ne demande pas la consécration de la vérité, ne demande pas à être suivi parce que vrai, n'impose rien et n'attend rien. Sa vraie destinée est l'errance et sa vraie conscience est celle de sa dissémination. Celui qui renonce à la prétention de vérité renonce en même temps à la volonté de dominer. Dominer signifie créer un espace de vérité dans lequel celui qui énonce la vérité est le maître légitime de ceux qui sont dans le champ de la conception du monde dont cette vérité est la pierre angulaire.

  16. eDT and Model-based Configuration of 12GeV CEBAF

    Energy Technology Data Exchange (ETDEWEB)

    Turner, Dennison L. [Jefferson Lab, Newport News, VA (United States)

    2015-09-01

    This poster will discuss model-driven setup of CEBAF for the 12GeV era, focusing on the elegant Download Tool (eDT). eDT is a new operator tool that generates magnet design setpoints for various machine energies and pass configurations. eDT was developed in the effort towards a process for reducing machine configuration time and reproducibility by way of an accurate accelerator model.

  17. Outlooks for mathematical modelling of the glass melting process

    Energy Technology Data Exchange (ETDEWEB)

    Waal, H. de [TNO Institute of Applied Physics, Delft (Netherlands)

    1997-12-31

    Mathematical modelling is nowadays a standard tool for major producers of float glass, T.V. glass and fiberglass. Also for container glass furnaces, glass tank modelling proves to be a valuable method to optimize process conditions. Mathematical modelling is no longer just a way to visualize the flow patterns and to provide data on heat transfer. It can also predict glass quality in relation to process parameters, because all chemical and physical phenomena are included in the latest generation of models, based on experimental and theoretical research on these phenomena.

  18. Modeling and analysis of micro-WEDM process of titanium alloy (Ti–6Al–4V using response surface approach

    Directory of Open Access Journals (Sweden)

    P. Sivaprakasam

    2014-12-01

    Full Text Available Micro-machining technology is effectively used in modern manufacturing industries. This paper investigates the influence of three different input parameters such as voltage, capacitance and feed rate of micro-wire electrical discharge machining (micro-WEDM performances of material removal rate (MRR, Kerf width (KW and surface roughness (SR using response surface methodology with central composite design (CCD. The experiments are carried out on titanium alloy (Ti–6Al–4V. The machining characteristics are significantly influenced by the electrical and non-electrical parameters in micro-WEDM process. Analysis of variance (ANOVA was performed to find out the significant influence of each factor. The model developed can use a genetic algorithm (GA to determine the optimal machining conditions using multi-objective optimization technique. The optimal machining performance of material removal rate, Kerf width and surface roughness are 0.01802 mm3/min, 101.5 μm and 0.789 μm, respectively, using this optimal machining conditions viz. voltage 100 V, capacitance 10 nF and feed rate 15 μm/s.

  19. Ultrasonic-assisted manufacturing processes: Variational model and numerical simulations

    KAUST Repository

    Siddiq, Amir

    2012-04-01

    We present a computational study of ultrasonic assisted manufacturing processes including sheet metal forming, upsetting, and wire drawing. A fully variational porous plasticity model is modified to include ultrasonic softening effects and then utilized to account for instantaneous softening when ultrasonic energy is applied during deformation. Material model parameters are identified via inverse modeling, i.e. by using experimental data. The versatility and predictive ability of the model are demonstrated and the effect of ultrasonic intensity on the manufacturing process at hand is investigated and compared qualitatively with experimental results reported in the literature. © 2011 Elsevier B.V. All rights reserved.

  20. SpaceCube v2.0 Space Flight Hybrid Reconfigurable Data Processing System

    Science.gov (United States)

    Petrick, Dave

    2014-01-01

    This paper details the design architecture, design methodology, and the advantages of the SpaceCube v2.0 high performance data processing system for space applications. The purpose in building the SpaceCube v2.0 system is to create a superior high performance, reconfigurable, hybrid data processing system that can be used in a multitude of applications including those that require a radiation hardened and reliable solution. The SpaceCube v2.0 system leverages seven years of board design, avionics systems design, and space flight application experiences. This paper shows how SpaceCube v2.0 solves the increasing computing demands of space data processing applications that cannot be attained with a standalone processor approach.The main objective during the design stage is to find a good system balance between power, size, reliability, cost, and data processing capability. These design variables directly impact each other, and it is important to understand how to achieve a suitable balance. This paper will detail how these critical design factors were managed including the construction of an Engineering Model for an experiment on the International Space Station to test out design concepts. We will describe the designs for the processor card, power card, backplane, and a mission unique interface card. The mechanical design for the box will also be detailed since it is critical in meeting the stringent thermal and structural requirements imposed by the processing system. In addition, the mechanical design uses advanced thermal conduction techniques to solve the internal thermal challenges.The SpaceCube v2.0 processing system is based on an extended version of the 3U cPCI standard form factor where each card is 190mm x 100mm in size The typical power draw of the processor card is 8 to 10W and scales with application complexity. The SpaceCube v2.0 data processing card features two Xilinx Virtex-5 QV Field Programmable Gate Arrays (FPGA), eight memory modules, a monitor

  1. Residence time modeling of hot melt extrusion processes.

    Science.gov (United States)

    Reitz, Elena; Podhaisky, Helmut; Ely, David; Thommes, Markus

    2013-11-01

    The hot melt extrusion process is a widespread technique to mix viscous melts. The residence time of material in the process frequently determines the product properties. An experimental setup and a corresponding mathematical model were developed to evaluate residence time and residence time distribution in twin screw extrusion processes. The extrusion process was modeled as the convolution of a mass transport process described by a Gaussian probability function, and a mixing process represented by an exponential function. The residence time of the extrusion process was determined by introducing a tracer at the extruder inlet and measuring the tracer concentration at the die. These concentrations were fitted to the residence time model, and an adequate correlation was found. Different parameters were derived to characterize the extrusion process including the dead time, the apparent mixing volume, and a transport related axial mixing. A 2(3) design of experiments was performed to evaluate the effect of powder feed rate, screw speed, and melt viscosity of the material on the residence time. All three parameters affect the residence time of material in the extruder. In conclusion, a residence time model was developed to interpret experimental data and to get insights into the hot melt extrusion process. Copyright © 2013 Elsevier B.V. All rights reserved.

  2. Similarities and Differences in Mechanical Alloying Processes of V-Si-B and Mo-Si-B Powders

    Directory of Open Access Journals (Sweden)

    Manja Krüger

    2016-10-01

    Full Text Available V-Si-B and Mo-Si-B alloys are currently the focus of materials research due to their excellent high temperature capabilities. To optimize the mechanical alloying (MA process for these materials, we compare microstructures, morphology and particles size as well as hardness evolution during the milling process for the model alloys V-9Si-13B and Mo-9Si-8B. A variation of the rotational speed of the planetary ball mill and the type of grinding materials is therefore investigated. These modifications result in different impact energies during ball-powder-wall collisions, which are quantitatively described in this comparative study. Processing with tungsten carbide vials and balls provides slightly improved impact energies compared to vials and balls made of steel. However, contamination of the mechanically alloyed powders with flaked particles of tungsten carbide is unavoidable. In the case of using steel grinding materials, Fe contaminations are also detectable, which are solved in the V and Mo solid solution phases, respectively. Typical mechanisms that occur during the MA process such as fracturing and comminution are analyzed using the comminution rate KP. In both alloys, the welding processes are more pronounced compared to the fracturing processes.

  3. An analytical turn-on power loss model for 650-V GaN eHEMTs

    DEFF Research Database (Denmark)

    Shen, Yanfeng; Wang, Huai; Shen, Zhan

    2018-01-01

    This paper proposes an improved analytical turn-on power loss model for 650-V GaN eHEMTs. The static characteristics, i.e., the parasitic capacitances and transconductance, are firstly modeled. Then the turn-on process is divided into multiple stages and analyzed in detail; as results, the time-d...

  4. Process modelling on a canonical basis[Process modelling; Canonical modelling

    Energy Technology Data Exchange (ETDEWEB)

    Siepmann, Volker

    2006-12-20

    Based on an equation oriented solving strategy, this thesis investigates a new approach to process modelling. Homogeneous thermodynamic state functions represent consistent mathematical models of thermodynamic properties. Such state functions of solely extensive canonical state variables are the basis of this work, as they are natural objective functions in optimisation nodes to calculate thermodynamic equilibrium regarding phase-interaction and chemical reactions. Analytical state function derivatives are utilised within the solution process as well as interpreted as physical properties. By this approach, only a limited range of imaginable process constraints are considered, namely linear balance equations of state variables. A second-order update of source contributions to these balance equations is obtained by an additional constitutive equation system. These equations are general dependent on state variables and first-order sensitivities, and cover therefore practically all potential process constraints. Symbolic computation technology efficiently provides sparsity and derivative information of active equations to avoid performance problems regarding robustness and computational effort. A benefit of detaching the constitutive equation system is that the structure of the main equation system remains unaffected by these constraints, and a priori information allows to implement an efficient solving strategy and a concise error diagnosis. A tailor-made linear algebra library handles the sparse recursive block structures efficiently. The optimisation principle for single modules of thermodynamic equilibrium is extended to host entire process models. State variables of different modules interact through balance equations, representing material flows from one module to the other. To account for reusability and encapsulation of process module details, modular process modelling is supported by a recursive module structure. The second-order solving algorithm makes it

  5. Influence of ∼7 keV sterile neutrino dark matter on the process of reionization

    International Nuclear Information System (INIS)

    Rudakovskyi, Anton; Iakubovskyi, Dmytro

    2016-01-01

    Recent reports of a weak unidentified emission line at ∼3.5 keV found in spectra of several matter-dominated objects may give a clue to resolve the long-standing problem of dark matter. One of the best physically motivated particle candidate able to produce such an extra line is sterile neutrino with the mass of ∼7 keV . Previous works show that sterile neutrino dark matter with parameters consistent with the new line measurement modestly affects structure formation compared to conventional cold dark matter scenario. In this work, we concentrate for the first time on contribution of the sterile neutrino dark matter able to produce the observed line at ∼3.5 keV, to the process of reionization. By incorporating dark matter power spectra for ∼7 keV sterile neutrinos into extended semi-analytical 'bubble' model of reionization we obtain that such sterile neutrino dark matter would produce significantly sharper reionization compared to widely used cold dark matter models, impossible to 'imitate' within the cold dark matter scenario under any reasonable choice of our model parameters, and would have a clear tendency of lowering both the redshift of reionization and the electron scattering optical depth (although the difference is still below the existing model uncertainties). Further dedicated studies of reionization (such as 21 cm measurements or studies of kinetic Sunyaev-Zeldovich effect) will thus be essential for reconstruction of particle candidate responsible the ∼3.5 keV line.

  6. Development of a Process for Separation of Mogroside V from Siraitia grosvenorii by Macroporous Resins

    Directory of Open Access Journals (Sweden)

    Ping Hu

    2011-08-01

    Full Text Available A separation method was developed for the preparative separation and enrichment of the non-caloric sweetener mogroside V from Siraitia grosvenorii. The adsorption properties of six macroporous resins were evaluated. Results showed that HZ 806 resin offered the best adsorption and desorption capacities. Based on the adsorption experiments on HZ 806, the adsorption data were found to fit the Freundlich model well. The pseudo-second-order kinetic model showed the highest correlation with the experimental results. Separation was performed with deionized water and 40% aqueous ethanol solution as mobile phases. In a typical run, 100 g of herb was processed and 3.38 g of mogroside V with a purity of 10.7% was harvested. This separation method provided a 15.1-fold increase in the purification factor from 0.5% to 10.7%. The present study showed that HZ 806 resins were effective for the separation and enrichment of mogroside V from S. grosvenorii.

  7. Search for the Standard Model Higgs boson in $e^+ e^-$ collisions at $\\sqrt{s}$ up to 202 GeV

    CERN Document Server

    Acciarri, M.; Adriani, O.; Aguilar-Benitez, M.; Alcaraz, J.; Alemanni, G.; Allaby, J.; Aloisio, A.; Alviggi, M.G.; Ambrosi, G.; Anderhub, H.; Andreev, Valery P.; Angelescu, T.; Anselmo, F.; Arefev, A.; Azemoon, T.; Aziz, T.; Bagnaia, P.; Bajo, A.; Baksay, L.; Balandras, A.; Baldew, S.V.; Banerjee, S.; Banerjee, Sw.; Barczyk, A.; Barillere, R.; Bartalini, P.; Basile, M.; Batalova, N.; Battiston, R.; Bay, A.; Becattini, F.; Becker, U.; Behner, F.; Bellucci, L.; Berbeco, R.; Berdugo, J.; Berges, P.; Bertucci, B.; Betev, B.L.; Bhattacharya, S.; Biasini, M.; Biland, A.; Blaising, J.J.; Blyth, S.C.; Bobbink, G.J.; Bohm, A.; Boldizsar, L.; Borgia, B.; Bourilkov, D.; Bourquin, M.; Braccini, S.; Branson, J.G.; Brochu, F.; Buffini, A.; Buijs, A.; Burger, J.D.; Burger, W.J.; Cai, X.D.; Capell, M.; Cara Romeo, G.; Carlino, G.; Cartacci, A.M.; Casaus, J.; Castellini, G.; Cavallari, F.; Cavallo, N.; Cecchi, C.; Cerrada, M.; Cesaroni, F.; Chamizo, M.; Chang, Y.H.; Chaturvedi, U.K.; Chemarin, M.; Chen, A.; Chen, G.; Chen, G.M.; Chen, H.F.; Chen, H.S.; Chiefari, G.; Cifarelli, L.; Cindolo, F.; Civinini, C.; Clare, I.; Clare, R.; Coignet, G.; Colino, N.; Costantini, S.; Cotorobai, F.; de la Cruz, B.; Csilling, A.; Cucciarelli, S.; Dai, T.S.; van Dalen, J.A.; D'Alessandro, R.; de Asmundis, R.; Deglon, P.; Degre, A.; Deiters, K.; della Volpe, D.; Delmeire, E.; Denes, P.; DeNotaristefani, F.; De Salvo, A.; Diemoz, M.; Dierckxsens, M.; van Dierendonck, D.; Dionisi, C.; Dittmar, M.; Dominguez, A.; Doria, A.; Dova, M.T.; Duchesneau, D.; Dufournaud, D.; Duinker, P.; El Mamouni, H.; Engler, A.; Eppling, F.J.; Erne, F.C.; Ewers, A.; Extermann, P.; Fabre, M.; Falagan, M.A.; Falciano, S.; Favara, A.; Fay, J.; Fedin, O.; Felcini, M.; Ferguson, T.; Fesefeldt, H.; Fiandrini, E.; Field, J.H.; Filthaut, F.; Fisher, P.H.; Fisk, I.; Forconi, G.; Freudenreich, K.; Furetta, C.; Galaktionov, Iouri; Ganguli, S.N.; Garcia-Abia, Pablo; Gataullin, M.; Gau, S.S.; Gentile, S.; Gheordanescu, N.; Giagu, S.; Gong, Z.F.; Grenier, Gerald Jean; Grimm, O.; Gruenewald, M.W.; Guida, M.; van Gulik, R.; Gupta, V.K.; Gurtu, A.; Gutay, L.J.; Haas, D.; Hasan, A.; Hatzifotiadou, D.; Hebbeker, T.; Herve, Alain; Hidas, P.; Hirschfelder, J.; Hofer, H.; Holzner, G.; Hoorani, H.; Hou, S.R.; Hu, Y.; Iashvili, I.; Jin, B.N.; Jones, Lawrence W.; de Jong, P.; Josa-Mutuberria, I.; Khan, R.A.; Kafer, D.; Kaur, M.; Kienzle-Focacci, M.N.; Kim, D.; Kim, J.K.; Kirkby, Jasper; Kiss, D.; Kittel, W.; Klimentov, A.; Konig, A.C.; Kopal, M.; Kopp, A.; Koutsenko, V.; Kraber, M.; Kraemer, R.W.; Krenz, W.; Kruger, A.; Kunin, A.; Ladron de Guevara, P.; Laktineh, I.; Landi, G.; Lebeau, M.; Lebedev, A.; Lebrun, P.; Lecomte, P.; Lecoq, P.; Le Coultre, P.; Lee, H.J.; Le Goff, J.M.; Leiste, R.; Levtchenko, P.; Li, C.; Likhoded, S.; Lin, C.H.; Lin, W.T.; Linde, F.L.; Lista, L.; Liu, Z.A.; Lohmann, W.; Longo, E.; Lu, Y.S.; Lubelsmeyer, K.; Luci, C.; Luckey, David; Lugnier, L.; Luminari, L.; Lustermann, W.; Ma, W.G.; Maity, M.; Malgeri, L.; Malinin, A.; Mana, C.; Mangeol, D.; Mans, J.; Marian, G.; Martin, J.P.; Marzano, F.; Mazumdar, K.; McNeil, R.R.; Mele, S.; Merola, L.; Meschini, M.; Metzger, W.J.; von der Mey, M.; Mihul, A.; Milcent, H.; Mirabelli, G.; Mnich, J.; Mohanty, G.B.; Moulik, T.; Muanza, G.S.; Muijs, A.J.M.; Musicar, B.; Musy, M.; Napolitano, M.; Nessi-Tedaldi, F.; Newman, H.; Niessen, T.; Nisati, A.; Kluge, Hannelies; Ofierzynski, R.; Organtini, G.; Oulianov, A.; Palomares, C.; Pandoulas, D.; Paoletti, S.; Paolucci, P.; Paramatti, R.; Park, H.K.; Park, I.H.; Passaleva, G.; Patricelli, S.; Paul, Thomas Cantzon; Pauluzzi, M.; Paus, C.; Pauss, F.; Pedace, M.; Pensotti, S.; Perret-Gallix, D.; Petersen, B.; Piccolo, D.; Pierella, F.; Pieri, M.; Piroue, P.A.; Pistolesi, E.; Plyaskin, V.; Pohl, M.; Pojidaev, V.; Postema, H.; Pothier, J.; Prokofev, D.O.; Prokofiev, D.; Quartieri, J.; Rahal-Callot, G.; Rahaman, M.A.; Raics, P.; Raja, N.; Ramelli, R.; Rancoita, P.G.; Ranieri, R.; Raspereza, A.; Raven, G.; Razis, P.; Ren, D.; Rescigno, M.; Reucroft, S.; Riemann, S.; Riles, Keith; Rodin, J.; Roe, B.P.; Romero, L.; Rosca, A.; Rosier-Lees, S.; Roth, Stefan; Rosenbleck, C.; Roux, B.; Rubio, J.A.; Ruggiero, G.; Rykaczewski, H.; Saremi, S.; Sarkar, S.; Salicio, J.; Sanchez, E.; Sanders, M.P.; Schafer, C.; Schegelsky, V.; Schmidt-Kaerst, S.; Schmitz, D.; Schopper, H.; Schotanus, D.J.; Schwering, G.; Sciacca, C.; Seganti, A.; Servoli, L.; Shevchenko, S.; Shivarov, N.; Shoutko, V.; Shumilov, E.; Shvorob, A.; Siedenburg, T.; Son, D.; Smith, B.; Spillantini, P.; Steuer, M.; Stickland, D.P.; Stone, A.; Stoyanov, B.; Straessner, A.; Sudhakar, K.; Sultanov, G.; Sun, L.Z.; Sushkov, S.; Suter, H.; Swain, J.D.; Szillasi, Z.; Sztaricskai, T.; Tang, X.W.; Tauscher, L.; Taylor, L.; Tellili, B.; Teyssier, D.; Timmermans, Charles; Ting, Samuel C.C.; Ting, S.M.; Tonwar, S.C.; Toth, J.; Tully, C.; Tung, K.L.; Uchida, Y.; Ulbricht, J.; Valente, E.; Vesztergombi, G.; Vetlitsky, I.; Vicinanza, D.; Viertel, G.; Villa, S.; Vivargent, M.; Vlachos, S.; Vodopianov, I.; Vogel, H.; Vogt, H.; Vorobev, I.; Vorobov, A.A.; Vorvolakos, A.; Wadhwa, M.; Wallraff, W.; Wang, M.; Wang, X.L.; Wang, Z.M.; Weber, A.; Weber, M.; Wienemann, P.; Wilkens, H.; Wu, S.X.; Wynhoff, S.; Xia, L.; Xu, Z.Z.; Yamamoto, J.; Yang, B.Z.; Yang, C.G.; Yang, H.J.; Yang, M.; Ye, J.B.; Yeh, S.C.; Zalite, A.; Zalite, Yu.; Zhang, Z.P.; Zhu, G.Y.; Zhu, R.Y.; Zichichi, A.; Zilizi, G.; Zimmermann, B.; Zoller, M.

    2001-01-01

    The Standard Model Higgs boson is searched for in 233.2 pb-1 of data collected by the L3 detector at centre of mass energies from 192 GeV to 202 GeV. These data are consistent with the expectations of Standard Model processes and no evidence of a Higgs signal is observed. A lower limit on the mass of the Standard Model Higgs boson of 107.0 GeV is set at the 95% confidence level.

  8. The iFlow modelling framework v2.4: a modular idealized process-based model for flow and transport in estuaries

    Science.gov (United States)

    Dijkstra, Yoeri M.; Brouwer, Ronald L.; Schuttelaars, Henk M.; Schramkowski, George P.

    2017-07-01

    The iFlow modelling framework is a width-averaged model for the systematic analysis of the water motion and sediment transport processes in estuaries and tidal rivers. The distinctive solution method, a mathematical perturbation method, used in the model allows for identification of the effect of individual physical processes on the water motion and sediment transport and study of the sensitivity of these processes to model parameters. This distinction between processes provides a unique tool for interpreting and explaining hydrodynamic interactions and sediment trapping. iFlow also includes a large number of options to configure the model geometry and multiple choices of turbulence and salinity models. Additionally, the model contains auxiliary components, including one that facilitates easy and fast sensitivity studies. iFlow has a modular structure, which makes it easy to include, exclude or change individual model components, called modules. Depending on the required functionality for the application at hand, modules can be selected to construct anything from very simple quasi-linear models to rather complex models involving multiple non-linear interactions. This way, the model complexity can be adjusted to the application. Once the modules containing the required functionality are selected, the underlying model structure automatically ensures modules are called in the correct order. The model inserts iteration loops over groups of modules that are mutually dependent. iFlow also ensures a smooth coupling of modules using analytical and numerical solution methods. This way the model combines the speed and accuracy of analytical solutions with the versatility of numerical solution methods. In this paper we present the modular structure, solution method and two examples of the use of iFlow. In the examples we present two case studies, of the Yangtze and Scheldt rivers, demonstrating how iFlow facilitates the analysis of model results, the understanding of the

  9. DNA scanning mechanism of T4 endonuclease V. Effect of NaCl concentration on processive nicking activity

    International Nuclear Information System (INIS)

    Gruskin, E.A.; Lloyd, R.S.

    1986-01-01

    T4 endonuclease V is a pyrimidine dimer-specific endonuclease which generates incisions in DNA at the sites of pyrimidine dimers by a processive reaction mechanism. A model is presented in which the degree of processivity is directly related to the efficacy of the one-dimensional diffusion of endonuclease V on DNA by which the enzyme locates pyrimidine dimers. The modulation of the processive nicking activity of T4 endonuclease V on superhelical covalently closed circular DNA (form I) which contains pyrimidine dimers has been investigated as a function of the ionic strength of the reaction. Agarose gel electrophoresis was used to separate the three topological forms of the DNA which were generated in time course reactions of endonuclease V with dimer-containing form I DNA in the absence of NaCl, and in 25, 50, and 100 mM NaCl. The degree of processivity was evaluated in terms of the mass fraction of form III (linear) DNA which was produced as a function of the fraction of form I DNA remaining. Processivity is maximal in the absence of NaCl and decreases as the NaCl concentration is increased. At 100 mM NaCl, processivity is abolished and endonuclease V generates incisions in DNA at the site of dimers by a distributive reaction mechanism. The change from the distributive to a processive reaction mechanism occurs at NaCl concentrations slightly below 50 mM. The high degree of processivity which is observed in the absence of NaCl is reversible to the distributive mechanism, as demonstrated by experiments in which the NaCl concentration was increased during the time course reaction. In addition, unirradiated DNA inhibited the incision of irradiated DNA only at NaCl concentrations at which processivity was observed

  10. Assessing Motivation to Improve Learning: Practical Applications of Keller's MVP Model and ARCS-V Design Process

    Science.gov (United States)

    Angelo, Thomas A.

    2017-01-01

    This chapter applies John Keller's MVP model and, specifically, adapts the ARCS-V components of that model--defined and described in Chapter 1 of this issue of "New Directions for Teaching and Learning"--as a frame for exploring practical, research-based assessment, and feedback strategies and tools teachers can use to help students…

  11. CSRL-V: processed ENDF/B-V 227-neutron-group and pointwise cross-section libraries for criticality safety, reactor, and shielding studies

    International Nuclear Information System (INIS)

    Ford, W.E. III; Diggs, B.R.; Petrie, L.M.; Webster, C.C.; Westfall, R.M.

    1982-01-01

    A P 3 227-neutron-group cross-section library has been processed for the subsequent generation of problem-dependent fine- or broad-group cross sections for a broad range of applications, including shipping cask calculations, general criticality safety analyses, and reactor core and shielding analyses. The energy group structure covers the range 10 -5 eV - 20 MeV, including 79 thermal groups below 3 eV. The 129-material library includes processed data for all materials in the ENDF/B-V General Purpose File, several data sets prepared from LENDL data, hydrogen with water- and polyethyelene-bound thermal kernels, deuterium with C 2 O-bound thermal kernels, carbon with a graphite thermal kernel, a special 1/V data set, and a dose factor data set. The library, which is in AMPX master format, is designated CSRL-V (Criticality Safety Reference Library based on ENDF/B-V data). Also included in CSRL-V is a pointwise total, fission, elastic scattering, and (n,γ) cross-section library containing data sets for all ENDF/B-V resonance materials. Data in the pointwise library were processed with the infinite dilute approximation at a temperature of 296 0 K

  12. The iFlow modelling framework v2.4: a modular idealized process-based model for flow and transport in estuaries

    Directory of Open Access Journals (Sweden)

    Y. M. Dijkstra

    2017-07-01

    Full Text Available The iFlow modelling framework is a width-averaged model for the systematic analysis of the water motion and sediment transport processes in estuaries and tidal rivers. The distinctive solution method, a mathematical perturbation method, used in the model allows for identification of the effect of individual physical processes on the water motion and sediment transport and study of the sensitivity of these processes to model parameters. This distinction between processes provides a unique tool for interpreting and explaining hydrodynamic interactions and sediment trapping. iFlow also includes a large number of options to configure the model geometry and multiple choices of turbulence and salinity models. Additionally, the model contains auxiliary components, including one that facilitates easy and fast sensitivity studies. iFlow has a modular structure, which makes it easy to include, exclude or change individual model components, called modules. Depending on the required functionality for the application at hand, modules can be selected to construct anything from very simple quasi-linear models to rather complex models involving multiple non-linear interactions. This way, the model complexity can be adjusted to the application. Once the modules containing the required functionality are selected, the underlying model structure automatically ensures modules are called in the correct order. The model inserts iteration loops over groups of modules that are mutually dependent. iFlow also ensures a smooth coupling of modules using analytical and numerical solution methods. This way the model combines the speed and accuracy of analytical solutions with the versatility of numerical solution methods. In this paper we present the modular structure, solution method and two examples of the use of iFlow. In the examples we present two case studies, of the Yangtze and Scheldt rivers, demonstrating how iFlow facilitates the analysis of model results, the

  13. The iFlow modelling framework v2.4 : A modular idealized process-based model for flow and transport in estuaries

    NARCIS (Netherlands)

    Dijkstra, Y.M.; Brouwer, R.L.; Schuttelaars, H.M.; Schramkowski, G.P.

    2017-01-01

    The iFlow modelling framework is a width-averaged model for the systematic analysis of the water motion and sediment transport processes in estuaries and tidal rivers. The distinctive solution method, a mathematical perturbation method, used in the model allows for identification of the effect of

  14. Modeling styles in business process modeling

    NARCIS (Netherlands)

    Pinggera, J.; Soffer, P.; Zugal, S.; Weber, B.; Weidlich, M.; Fahland, D.; Reijers, H.A.; Mendling, J.; Bider, I.; Halpin, T.; Krogstie, J.; Nurcan, S.; Proper, E.; Schmidt, R.; Soffer, P.; Wrycza, S.

    2012-01-01

    Research on quality issues of business process models has recently begun to explore the process of creating process models. As a consequence, the question arises whether different ways of creating process models exist. In this vein, we observed 115 students engaged in the act of modeling, recording

  15. Assessment of realizability constraints in v2-f turbulence models

    International Nuclear Information System (INIS)

    Sveningsson, A.; Davidson, L.

    2004-01-01

    The use of the realizability constraint in v 2 -f turbulence models is assessed by computing a stator vane passage flow. In this flow the stagnation region is large and it is shown that the time scale bound suggested by [Int. J. Heat Fluid Flow 17 (1995) 89] is well suited to prevent unphysical growth of turbulence kinetic energy. However, this constraint causes numerical instabilities when used in the equation for the relaxation parameter, f. It is also shown that the standard use of the realizability constraint in the v 2 -f model is inconsistent and some modifications are suggested. These changes of the v 2 -f model are examined and shown to have negligible effect on the overall performance of the v 2 -f model. In this work two different versions of the v 2 -f model are investigated and the results obtained are compared with experimental data. The model on a form similar to that originally suggested by Durbin (e.g. [AIAA J. 33 (1995) 659]) produced the overall best agreement with stator vane heat transfer data

  16. Statistically Modeling I-V Characteristics of CNT-FET with LASSO

    Science.gov (United States)

    Ma, Dongsheng; Ye, Zuochang; Wang, Yan

    2017-08-01

    With the advent of internet of things (IOT), the need for studying new material and devices for various applications is increasing. Traditionally we build compact models for transistors on the basis of physics. But physical models are expensive and need a very long time to adjust for non-ideal effects. As the vision for the application of many novel devices is not certain or the manufacture process is not mature, deriving generalized accurate physical models for such devices is very strenuous, whereas statistical modeling is becoming a potential method because of its data oriented property and fast implementation. In this paper, one classical statistical regression method, LASSO, is used to model the I-V characteristics of CNT-FET and a pseudo-PMOS inverter simulation based on the trained model is implemented in Cadence. The normalized relative mean square prediction error of the trained model versus experiment sample data and the simulation results show that the model is acceptable for digital circuit static simulation. And such modeling methodology can extend to general devices.

  17. Process modeling style

    CERN Document Server

    Long, John

    2014-01-01

    Process Modeling Style focuses on other aspects of process modeling beyond notation that are very important to practitioners. Many people who model processes focus on the specific notation used to create their drawings. While that is important, there are many other aspects to modeling, such as naming, creating identifiers, descriptions, interfaces, patterns, and creating useful process documentation. Experience author John Long focuses on those non-notational aspects of modeling, which practitioners will find invaluable. Gives solid advice for creating roles, work produ

  18. Excimer laser processing of Ti-6Al-4V

    International Nuclear Information System (INIS)

    Folkes, J.A.; Shibata, K.

    1994-01-01

    The effect of the excimer laser on the surface of Ti-6Al-4V is reported. Particular concentration is given to surface modification for potential materials processing applications. Results showed that: (1) there is an optimum energy for smoothing titanium; (2) at this energy density increasing the number of pulses has some, but not a significant, effect on the smoothing process; and (3) relatively smooth surfaces could be achieved at higher energy densities if the sample was processed in a helium atmosphere. Other typical surface modifications and features are also reported, including the effect of different gases on the process

  19. Processive nicking activity of T4 endonuclease V on UV-irradiated chromatin

    International Nuclear Information System (INIS)

    Gruskin, E.A.; Lloyd, R.S.

    1986-01-01

    T4 endonuclease V initiates the excision repair of pyrimidine dimers in UV-irradiated T4 infected E. coli cells. The pyrimidine dimer specific nicking activity of T4 endonuclease V functions by a processive scanning on UV-irradiated DNA. Previously it has been demonstrated that introduction of endonuclease V into repair-deficient human cells causes a restoration of UV survival in these cells. This demonstrates that endonuclease V is competent to incise mammalian DNA at the site of pyrimidine dimers. In order to assess the ability of endonuclease V to act processively on DNA associated as chromatin, minichromosomes were prepared for use as a substrate. Form I DNA was reconstituted with H3, H4 +/- H1 histones by sequential dialysis steps from 2.0 M NaCl to 50 mM NaCl. Time course reactions were performed with minichromosomes containing 10 and 25 dimers per molecule. In each case the rate of disappearance of form I DNA which was associated as chromatin was decreased relative to that of naked form I DNA. Concurrent with that observation, the rate and extent of appearance of form III DNA was increased with the DNA in minichromosomes relative to naked DNA. This is diagnostic of an enhancement of processivity. The inclusion of H1 in the minichromosomes resulted in a slight additional increase in processivity relative to minichromosomes consisting only of H3 and H4

  20. Transforming Collaborative Process Models into Interface Process Models by Applying an MDA Approach

    Science.gov (United States)

    Lazarte, Ivanna M.; Chiotti, Omar; Villarreal, Pablo D.

    Collaborative business models among enterprises require defining collaborative business processes. Enterprises implement B2B collaborations to execute these processes. In B2B collaborations the integration and interoperability of processes and systems of the enterprises are required to support the execution of collaborative processes. From a collaborative process model, which describes the global view of the enterprise interactions, each enterprise must define the interface process that represents the role it performs in the collaborative process in order to implement the process in a Business Process Management System. Hence, in this work we propose a method for the automatic generation of the interface process model of each enterprise from a collaborative process model. This method is based on a Model-Driven Architecture to transform collaborative process models into interface process models. By applying this method, interface processes are guaranteed to be interoperable and defined according to a collaborative process.

  1. Angular distribution of Drell-Yan process at hadron colliders to NLO-QCD in models of TeV scale gravity

    International Nuclear Information System (INIS)

    Mathews, Prakash; Ravindran, V.

    2006-01-01

    In TeV scale gravity models, for dilepton production at hadron colliders, we present the NLO-QCD corrections for the double differential cross section in the invariant mass and scattering angle. For both ADD and RS models, the quantitative impact of QCD corrections for extra dimension searches at LHC and Tevatron are investigated. We present the K-factors for both ADD and RS models at LHC and Tevatron. Inclusion of QCD corrections to NLO stabilises the cross section with respect to scale variations

  2. Exclusive processes at JLab at 6 GeV

    Directory of Open Access Journals (Sweden)

    Kim Andrey

    2015-01-01

    Full Text Available Deeply virtual exclusive reactions provide a unique opportunity to probe the complex internal structure of the nucleon. They allow to access information about the correlations between parton transverse spatial and longitudinal momentum distributions from experimental observables. Dedicated experiments to study Deeply Virtual Compton Scattering (DVCS and Deeply Virtual Meson Production (DVMP have been carried out at Jefferson Lab using continuous electron beam with energies up to 6 GeV. Unpolarized cross sections, beam, target and double spin asymmetries have been measured for DVCS as well as for π0 exclusive electroproduction. The data from Hall B provide a wide kinematic coverage with Q2=1-4.5 GeV2, xB=0.1-0.5, and −t up to 2 GeV2. Hall A data have limited kinematic range partially overlapping with Hall B kinematics but provide a high accuracy measurements. Scaling tests of the DVCS cross sections provide solid evidence of twist-2 dominance, which makes chiral-even GPDs accessible even at modest Q2. We will discuss the interpretation of these data in terms of Generalized Parton Distributions (GPDs model. Successful description of the recent CLAS π0 exclusive production data within the framework of the GPD-based model provides a unique opportunity to access the chiral-odd GPDs.

  3. Flexible software process lines in practice: A metamodel-based approach to effectively construct and manage families of software process models

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Ternité, Thomas; Friedrich, Jan

    2017-01-01

    Process flexibility and adaptability is frequently discussed, and several proposals aim to improve software processes for a given organization-/project context. A software process line (SPrL) is an instrument to systematically construct and manage variable software processes, by combining pre-def......: A metamodel-based approach to effectively construct and manage families of software process models [Ku16]. This paper was published as original research article in the Journal of Systems and Software.......Process flexibility and adaptability is frequently discussed, and several proposals aim to improve software processes for a given organization-/project context. A software process line (SPrL) is an instrument to systematically construct and manage variable software processes, by combining pre...... to construct flexible SPrLs and show its practical application in the German V-Modell XT. We contribute a proven approach that is presented as metamodel fragment for reuse and implementation in further process modeling approaches. This summary refers to the paper Flexible software process lines in practice...

  4. Determination of optimum parameters of the technological process for plates forming from V95 and V-1461 alloys in creep applied in aircrafts constructed by “Sukhoi design bureau”

    Science.gov (United States)

    Raevskaya, G. A.; Zakharchenko, K.; Larichkin, A.

    2017-10-01

    The research is devoted to the scientific justification of metal processing by pressure with the help of thick monolithic plates forming (thickness 40 mm) from the V95 (analog 7475) (Al-Zn-Mg-Cu) and V-1461 (analog 2099) (Al-Cu-Li-Zn) alloys in creep and close-to-superplasticity. Optimum parameters of the technological process of plate forming are described. The effect of temperature on the magnitude of mechanical stresses (relaxation) during the tests of materials on pure bending is experimentally determined. Forming of thick plates (40 mm) on the UFP-1M unit, and the control of the obtained surface, in comparison with the given electronic model, made it possible to experimentally determine the time and number of forming stages. Mechanical properties of the material after the technological process and heat treatment are preliminary evaluated. The efficiency of using the obtained parameters of the technological process and treatment of metals by pressure in such methods in general is shown.

  5. Modelling of fluoride removal via batch monopolar electrocoagulation process using aluminium electrodes

    Science.gov (United States)

    Amri, N.; Hashim, M. I.; Ismail, N.; Rohman, F. S.; Bashah, N. A. A.

    2017-09-01

    Electrocoagulation (EC) is a promising technology that extensively used to remove fluoride ions efficiently from industrial wastewater. However, it has received very little consideration and understanding on mechanism and factors that affecting the fluoride removal process. In order to determine the efficiency of fluoride removal in EC process, the effect of operating parameters such as voltage and electrolysis time were investigated in this study. A batch experiment with monopolar aluminium electrodes was conducted to identify the model of fluoride removal using empirical model equation. The EC process was investigated using several parameters which include voltage (3 - 12 V) and electrolysis time (0 - 60 minutes) at a constant initial fluoride concentration of 25 mg/L. The result shows that the fluoride removal efficiency increased steadily with increasing voltage and electrolysis time. The best fluoride removal efficiency was obtained with 94.8 % removal at 25 mg/L initial fluoride concentration, voltage of 12 V and 60 minutes electrolysis time. The results indicated that the rate constant, k and number of order, n decreased as the voltage increased. The rate of fluoride removal model was developed based on the empirical model equation using the correlation of k and n. Overall, the result showed that EC process can be considered as a potential alternative technology for fluoride removal in wastewater.

  6. Rational Unified Process jako metodika vývoje softwaru

    OpenAIRE

    Rytíř, Vladimír

    2008-01-01

    Cílem této práce je seznámit se s procesem vývoje softwaru dle vybraných metodik s hlavním zaměřením na metodiku Rational Unified Process od firmy IBM. Incepční a elaborační fáze této metody také aplikuji na praktickém  příkladě. Goal of my work is to introduce software development process metods specialized to Rational Unified Process metod from IBM. I aplicate inception and elaboration phases of RUP on practical example. B

  7. The use and misuse of V(c,max) in Earth System Models.

    Science.gov (United States)

    Rogers, Alistair

    2014-02-01

    Earth System Models (ESMs) aim to project global change. Central to this aim is the need to accurately model global carbon fluxes. Photosynthetic carbon dioxide assimilation by the terrestrial biosphere is the largest of these fluxes, and in many ESMs is represented by the Farquhar, von Caemmerer and Berry (FvCB) model of photosynthesis. The maximum rate of carboxylation by the enzyme Rubisco, commonly termed V c,max, is a key parameter in the FvCB model. This study investigated the derivation of the values of V c,max used to represent different plant functional types (PFTs) in ESMs. Four methods for estimating V c,max were identified; (1) an empirical or (2) mechanistic relationship was used to relate V c,max to leaf N content, (3) V c,max was estimated using an approach based on the optimization of photosynthesis and respiration or (4) calibration of a user-defined V c,max to obtain a target model output. Despite representing the same PFTs, the land model components of ESMs were parameterized with a wide range of values for V c,max (-46 to +77% of the PFT mean). In many cases, parameterization was based on limited data sets and poorly defined coefficients that were used to adjust model parameters and set PFT-specific values for V c,max. Examination of the models that linked leaf N mechanistically to V c,max identified potential changes to fixed parameters that collectively would decrease V c,max by 31% in C3 plants and 11% in C4 plants. Plant trait data bases are now available that offer an excellent opportunity for models to update PFT-specific parameters used to estimate V c,max. However, data for parameterizing some PFTs, particularly those in the Tropics and the Arctic are either highly variable or largely absent.

  8. 'TeV Gamma-ray Crisis' and an Anisotropic Space Model

    OpenAIRE

    Cho, Gi-Chol; Kamoshita, Jun-ichi; Matsunaga, Mariko; Sugamoto, Akio; Watanabe, Isamu

    2004-01-01

    To solve the `TeV gamma crisis', we examine a model whose one spatial direction is discretized at a high energy scale. Assuming the standard extra-galactic IR photon distribution, we evaluate the mean free-path of a energetic photon which acquires an effective mass in the model. For a wide range of the value of the lattice energy scale between a few TeV and around $10^{10}$ GeV, the mean free-path of a TeV energy photon can be enlarged enough to solve the `crisis'. Taking into account the eff...

  9. Deuteron-deuteron scattering at 3.0, 3.4 and 3.7 GeV/c, theoretical interpretation in the Glauber model

    International Nuclear Information System (INIS)

    Ballot, J.L.; L'Huillier, M.; Combes, M.P.; Tatischeff, B.

    1984-01-01

    An analysis of deuteron-deuteron scattering data at 3.0, 3.4 and 3.7 GeV/c is given in the framework of the Glauber NN multiple scattering model. The model accounts qualitatively well for the larger momentum transfer data. The model cannot reproduce the observed experimental bump at small transfer in the region of missing mass 2GeV/c, because pionic processes have been considered. Nevertheless it shows that the nucleon background is important

  10. Characterization of 10 MeV electron linac for radiation processing

    International Nuclear Information System (INIS)

    Petwal, V.C.; Rao, J.N.; Kaul, A.; Bapna, S.C.; Mulchandani, J.K.; Wanmode, Y.; Pandiyar, M.; Srivastava, P.; Jain, Akhilesh; Hanurkar, P.R.

    2006-01-01

    A radiation processing facility based on a 10 MeV LINAC is being set-up at RRCAT. In the course of commissioning various experiments have been carried-out to characterize the radiation field generated by the accelerator and subsequently to derive the operating parameters of the facility for radiation processing of various items. Results of the experiments are presented in the paper. (author)

  11. Performance of GeantV EM Physics Models

    Energy Technology Data Exchange (ETDEWEB)

    Amadio, G.; et al.

    2016-10-14

    The recent progress in parallel hardware architectures with deeper vector pipelines or many-cores technologies brings opportunities for HEP experiments to take advantage of SIMD and SIMT computing models. Launched in 2013, the GeantV project studies performance gains in propagating multiple particles in parallel, improving instruction throughput and data locality in HEP event simulation on modern parallel hardware architecture. Due to the complexity of geometry description and physics algorithms of a typical HEP application, performance analysis is indispensable in identifying factors limiting parallel execution. In this report, we will present design considerations and preliminary computing performance of GeantV physics models on coprocessors (Intel Xeon Phi and NVidia GPUs) as well as on mainstream CPUs.

  12. Performance of GeantV EM Physics Models

    Science.gov (United States)

    Amadio, G.; Ananya, A.; Apostolakis, J.; Aurora, A.; Bandieramonte, M.; Bhattacharyya, A.; Bianchini, C.; Brun, R.; Canal, P.; Carminati, F.; Cosmo, G.; Duhem, L.; Elvira, D.; Folger, G.; Gheata, A.; Gheata, M.; Goulas, I.; Iope, R.; Jun, S. Y.; Lima, G.; Mohanty, A.; Nikitina, T.; Novak, M.; Pokorski, W.; Ribon, A.; Seghal, R.; Shadura, O.; Vallecorsa, S.; Wenzel, S.; Zhang, Y.

    2017-10-01

    The recent progress in parallel hardware architectures with deeper vector pipelines or many-cores technologies brings opportunities for HEP experiments to take advantage of SIMD and SIMT computing models. Launched in 2013, the GeantV project studies performance gains in propagating multiple particles in parallel, improving instruction throughput and data locality in HEP event simulation on modern parallel hardware architecture. Due to the complexity of geometry description and physics algorithms of a typical HEP application, performance analysis is indispensable in identifying factors limiting parallel execution. In this report, we will present design considerations and preliminary computing performance of GeantV physics models on coprocessors (Intel Xeon Phi and NVidia GPUs) as well as on mainstream CPUs.

  13. Performance of GeantV EM Physics Models

    CERN Document Server

    Amadio, G; Apostolakis, J; Aurora, A; Bandieramonte, M; Bhattacharyya, A; Bianchini, C; Brun, R; Canal P; Carminati, F; Cosmo, G; Duhem, L; Elvira, D; Folger, G; Gheata, A; Gheata, M; Goulas, I; Iope, R; Jun, S Y; Lima, G; Mohanty, A; Nikitina, T; Novak, M; Pokorski, W; Ribon, A; Seghal, R; Shadura, O; Vallecorsa, S; Wenzel, S; Zhang, Y

    2017-01-01

    The recent progress in parallel hardware architectures with deeper vector pipelines or many-cores technologies brings opportunities for HEP experiments to take advantage of SIMD and SIMT computing models. Launched in 2013, the GeantV project studies performance gains in propagating multiple particles in parallel, improving instruction throughput and data locality in HEP event simulation on modern parallel hardware architecture. Due to the complexity of geometry description and physics algorithms of a typical HEP application, performance analysis is indispensable in identifying factors limiting parallel execution. In this report, we will present design considerations and preliminary computing performance of GeantV physics models on coprocessors (Intel Xeon Phi and NVidia GPUs) as well as on mainstream CPUs.

  14. Process, cost modeling and simulations for integrated project development of biomass for fuel and protein

    International Nuclear Information System (INIS)

    Pannir Selvam, P.V.; Wolff, D.M.B.; Souza Melo, H.N.

    1998-01-01

    The construction of the models for biomass project development are described. These models, first constructed using QPRO electronic spread sheet for Windows, are now being developed with the aid of visual and object oriented program as tools using DELPHI V.1 for windows and process simulator SUPERPRO, V.2.7 Intelligent Inc. These models render the process development problems with economic objectives to be solved very rapidly. The preliminary analysis of cost and investments of biomass utilisation projects which are included for this study are: steam, ammonia, carbon dioxide and alkali pretreatment process, methane gas production using anaerobic digestion process, aerobic composting, ethanol fermentation and distillation, effluent treatments using high rate algae production as well as cogeneration of energy for drying. The main project under developments are the biomass valuation projects with the elephant (Napier) grass, sugar cane bagasse and microalgae, using models for mass balance, equipment and production cost. The sensibility analyses are carried out to account for stochastic variation of the process yield, production volume, price variations, using Monte Carlo method. These models allow the identification of economical and scale up problems of the technology. The results obtained with few preliminary project development with few case studies are reported for integrated project development for fuel and protein using process and cost simulation models. (author)

  15. A spatially explicit hydro-ecological modeling framework (BEPS-TerrainLab V2.0): Model description and test in a boreal ecosystem in Eastern North America

    Science.gov (United States)

    Govind, Ajit; Chen, Jing Ming; Margolis, Hank; Ju, Weimin; Sonnentag, Oliver; Giasson, Marc-André

    2009-04-01

    SummaryA spatially explicit, process-based hydro-ecological model, BEPS-TerrainLab V2.0, was developed to improve the representation of ecophysiological, hydro-ecological and biogeochemical processes of boreal ecosystems in a tightly coupled manner. Several processes unique to boreal ecosystems were implemented including the sub-surface lateral water fluxes, stratification of vegetation into distinct layers for explicit ecophysiological representation, inclusion of novel spatial upscaling strategies and biogeochemical processes. To account for preferential water fluxes common in humid boreal ecosystems, a novel scheme was introduced based on laboratory analyses. Leaf-scale ecophysiological processes were upscaled to canopy-scale by explicitly considering leaf physiological conditions as affected by light and water stress. The modified model was tested with 2 years of continuous measurements taken at the Eastern Old Black Spruce Site of the Fluxnet-Canada Research Network located in a humid boreal watershed in eastern Canada. Comparison of the simulated and measured ET, water-table depth (WTD), volumetric soil water content (VSWC) and gross primary productivity (GPP) revealed that BEPS-TerrainLab V2.0 simulates hydro-ecological processes with reasonable accuracy. The model was able to explain 83% of the ET, 92% of the GPP variability and 72% of the WTD dynamics. The model suggests that in humid ecosystems such as eastern North American boreal watersheds, topographically driven sub-surface baseflow is the main mechanism of soil water partitioning which significantly affects the local-scale hydrological conditions.

  16. Visualizing the process of process modeling with PPMCharts

    NARCIS (Netherlands)

    Claes, J.; Vanderfeesten, I.T.P.; Pinggera, J.; Reijers, H.A.; Weber, B.; Poels, G.; La Rosa, M.; Soffer, P.

    2013-01-01

    In the quest for knowledge about how to make good process models, recent research focus is shifting from studying the quality of process models to studying the process of process modeling (often abbreviated as PPM) itself. This paper reports on our efforts to visualize this specific process in such

  17. Process modeling and control applied to real-time monitoring of distillation processes by near-infrared spectroscopy.

    Science.gov (United States)

    de Oliveira, Rodrigo R; Pedroza, Ricardo H P; Sousa, A O; Lima, Kássio M G; de Juan, Anna

    2017-09-08

    A distillation device that acquires continuous and synchronized measurements of temperature, percentage of distilled fraction and NIR spectra has been designed for real-time monitoring of distillation processes. As a process model, synthetic commercial gasoline batches produced in Brazil, which contain mixtures of pure gasoline blended with ethanol have been analyzed. The information provided by this device, i.e., distillation curves and NIR spectra, has served as initial information for the proposal of new strategies of process modeling and multivariate statistical process control (MSPC). Process modeling based on PCA batch analysis provided global distillation trajectories, whereas multiset MCR-ALS analysis is proposed to obtain a component-wise characterization of the distillation evolution and distilled fractions. Distillation curves, NIR spectra or compressed NIR information under the form of PCA scores and MCR-ALS concentration profiles were tested as the seed information to build MSPC models. New on-line PCA-based MSPC approaches, some inspired on local rank exploratory methods for process analysis, are proposed and work as follows: a) MSPC based on individual process observation models, where multiple local PCA models are built considering the sole information in each observation point; b) Fixed Size Moving Window - MSPC, in which local PCA models are built considering a moving window of the current and few past observation points; and c) Evolving MSPC, where local PCA models are built with an increasing window of observations covering all points since the beginning of the process until the current observation. Performance of different approaches has been assessed in terms of sensitivity to fault detection and number of false alarms. The outcome of this work will be of general use to define strategies for on-line process monitoring and control and, in a more specific way, to improve quality control of petroleum derived fuels and other substances submitted

  18. Standard model beyond the TeV

    International Nuclear Information System (INIS)

    Aurenche, P.

    1987-01-01

    The phenomenology of the standard model in the hadronic reactions in the 10 TeV range is described. The predictions of the model concerning the hadronic cross sections being based on the parton model, we first discuss the behaviour of the structure functions at the low values of X (x > 10 -4 ) which are attained at these energies and we show that the development of the leading logarithms equations allow us to calculate them. The production of W, Z, and gauge bosons and gauge boson pairs are reviewed. The Higgs boson production is discussed in detail according to his mass value [fr

  19. Recurrent V1-V2 interaction in early visual boundary processing.

    Science.gov (United States)

    Neumann, H; Sepp, W

    1999-11-01

    A majority of cortical areas are connected via feedforward and feedback fiber projections. In feedforward pathways we mainly observe stages of feature detection and integration. The computational role of the descending pathways at different stages of processing remains mainly unknown. Based on empirical findings we suggest that the top-down feedback pathways subserve a context-dependent gain control mechanism. We propose a new computational model for recurrent contour processing in which normalized activities of orientation selective contrast cells are fed forward to the next processing stage. There, the arrangement of input activation is matched against local patterns of contour shape. The resulting activities are subsequently fed back to the previous stage to locally enhance those initial measurements that are consistent with the top-down generated responses. In all, we suggest a computational theory for recurrent processing in the visual cortex in which the significance of local measurements is evaluated on the basis of a broader visual context that is represented in terms of contour code patterns. The model serves as a framework to link physiological with perceptual data gathered in psychophysical experiments. It handles a variety of perceptual phenomena, such as the local grouping of fragmented shape outline, texture surround and density effects, and the interpolation of illusory contours.

  20. Dynamic densification of metal matrix-coated fibre composites: modelling and processing

    International Nuclear Information System (INIS)

    Peng, H.X.; Dunne, F.P.E.; Grant, P.S.; Cantor, B.

    2005-01-01

    The consolidation processing of Ti-6Al-4V matrix-coated fibre (MCF) composite under vacuum hot pressing (VHP) has been investigated. A new test methodology has been developed for the determination of in situ matrix coating creep properties. In using the methodology, only a single, simple test is required, together with finite element modelling of the single fibre compression test. The creep coefficient and stress index have been determined for electron beam evaporated physical vapour deposited Ti-6Al-4V at 900 deg. C to be 1.23 x 10 -5 and 1.3, respectively. Consolidation experiments have been carried out on multi-ply MCF arrays under vacuum hot pressing. Finite element models have been developed for the dynamic consolidation of both square and hexagonal fibre packings. The creep constants for the Ti-6Al-4V, determined using the single fibre test, were assigned to the coating in the finite element models. Excellent agreement between predicted and experimental results was achieved, providing verification of the single fibre test methodology for the determination of creep constants

  1. Exploring the influence of constitutive models and associated parameters for the orthogonal machining of Ti6Al4V

    Science.gov (United States)

    Pervaiz, S.; Anwar, S.; Kannan, S.; Almarfadi, A.

    2018-04-01

    Ti6Al4V is known as difficult-to-cut material due to its inherent properties such as high hot hardness, low thermal conductivity and high chemical reactivity. Though, Ti6Al4V is utilized by industrial sectors such as aeronautics, energy generation, petrochemical and bio-medical etc. For the metal cutting community, competent and cost-effective machining of Ti6Al4V is a challenging task. To optimize cost and machining performance for the machining of Ti6Al4V, finite element based cutting simulation can be a very useful tool. The aim of this paper is to develop a finite element machining model for the simulation of Ti6Al4V machining process. The study incorporates material constitutive models namely Power Law (PL) and Johnson – Cook (JC) material models to mimic the mechanical behaviour of Ti6Al4V. The study investigates cutting temperatures, cutting forces, stresses, and plastic strains with respect to different PL and JC material models with associated parameters. In addition, the numerical study also integrates different cutting tool rake angles in the machining simulations. The simulated results will be beneficial to draw conclusions for improving the overall machining performance of Ti6Al4V.

  2. Validation of ASTEC V2 models for the behaviour of corium in the vessel lower head

    International Nuclear Information System (INIS)

    Carénini, L.; Fleurot, J.; Fichot, F.

    2014-01-01

    The paper is devoted to the presentation of validation cases carried out for the models describing the corium behaviour in the “lower plenum” of the reactor vessel implemented in the V2.0 version of the ASTEC integral code, jointly developed by IRSN (France) and GRS (Germany). In the ASTEC architecture, these models are grouped within the single ICARE module and they are all activated in typical accident scenarios. Therefore, it is important to check the validity of each individual model, as long as experiments are available for which a single physical process is involved. The results of ASTEC applications against the following experiments are presented: FARO (corium jet fragmentation), LIVE (heat transfer between a molten pool and the vessel), MASCA (separation and stratification of corium non miscible phases) and OLHF (mechanical failure of the vessel). Compared to the previous ASTEC V1.3 version, the validation matrix is extended. This work allows determining recommended values for some model parameters (e.g. debris particle size in the fragmentation model and criterion for debris bed liquefaction). Almost all the processes governing the corium behaviour, its thermal interaction with the vessel wall and the vessel failure are modelled in ASTEC and these models have been assessed individually with satisfactory results. The main uncertainties appear to be related to the calculation of transient evolutions

  3. ASTEC V2 severe accident integral code main features, current V2.0 modelling status, perspectives

    International Nuclear Information System (INIS)

    Chatelard, P.; Reinke, N.; Arndt, S.; Belon, S.; Cantrel, L.; Carenini, L.; Chevalier-Jabet, K.; Cousin, F.; Eckel, J.; Jacq, F.; Marchetto, C.; Mun, C.; Piar, L.

    2014-01-01

    The severe accident integral code ASTEC, jointly developed since almost 20 years by IRSN and GRS, simulates the behaviour of a whole nuclear power plant under severe accident conditions, including severe accident management by engineering systems and procedures. Since 2004, the ASTEC code is progressively becoming the reference European severe accident integral code through in particular the intensification of research activities carried out in the frame of the SARNET European network of excellence. The first version of the new series ASTEC V2 was released in 2009 to about 30 organizations worldwide and in particular to SARNET partners. With respect to the previous V1 series, this new V2 series includes advanced core degradation models (issued from the ICARE2 IRSN mechanistic code) and necessary extensions to be applicable to Gen. III reactor designs, notably a description of the core catcher component to simulate severe accidents transients applied to the EPR reactor. Besides these two key-evolutions, most of the other physical modules have also been improved and ASTEC V2 is now coupled to the SUNSET statistical tool to make easier the uncertainty and sensitivity analyses. The ASTEC models are today at the state of the art (in particular fission product models with respect to source term evaluation), except for quenching of a severely damage core. Beyond the need to develop an adequate model for the reflooding of a degraded core, the main other mean-term objectives are to further progress on the on-going extension of the scope of application to BWR and CANDU reactors, to spent fuel pool accidents as well as to accidents in both the ITER Fusion facility and Gen. IV reactors (in priority on sodium-cooled fast reactors) while making ASTEC evolving towards a severe accident simulator constitutes the main long-term objective. This paper presents the status of the ASTEC V2 versions, focussing on the description of V2.0 models for water-cooled nuclear plants

  4. SUSY simplified models at 14, 33, and 100 TeV proton colliders

    International Nuclear Information System (INIS)

    Cohen, Timothy; Golling, Tobias; Hance, Mike; Henrichs, Anna; Howe, Kiel; Loyal, Joshua; Padhi, Sanjay; Wacker, Jay G.

    2014-01-01

    Results are presented for a variety of SUSY Simplified Models at the 14 TeV LHC as well as a 33 and 100 TeV proton collider. Our focus is on models whose signals are driven by colored production. We present projections of the upper limit and discovery reach in the gluino-neutralino (for both light and heavy flavor decays), squark-neutralino, and gluino-squark Simplified Model planes. Depending on the model a jets + E T miss , mono-jet, or same-sign di-lepton search is applied. The impact of pileup is explored. This study utilizes the Snowmass backgrounds and combined detector. Assuming 3000/,fb −1 of integrated luminosity, a gluino that decays to light flavor quarks can be discovered below 2.3 TeV at the 14 TeV LHC and below 11 TeV at a 100 TeV machine

  5. Modelling, Analysis and Performance Evaluation of Power Conversion Unit in G2V/V2G Application—A Review

    Directory of Open Access Journals (Sweden)

    María Garcés Quílez

    2018-04-01

    Full Text Available In response to climate change, which is caused by the increasing pollution of the environment and leads to the deterioration of human health, future electricity generations should reduce reliance on fossil fuels by growing the use of clean and renewable energy generation sources and by using clean vehicle technologies. Battery storage systems have been recognized as one of the most promising approaches for supporting the renewable energy generation sources and cleanly powering vehicles instead of burning gasoline and diesel fuel. However, the cost of batteries is still a prominent barrier for their use in stationary and traction applications. As a rule, the cost of batteries can be decreased by lowering material costs, enhancing process efficiencies, and increasing production volume. Another more effective solution is called Vehicle-to-grid (V2G application. In V2G application, the battery system can be used to support the grid services, whereas the battery is still in the vehicle. To make a battery system economically viable for V2G/G2V applications, an effective power-electronics converter should be selected as well. This converter should be supported by an advanced control strategy. Therefore, this article provides a detailed technical assessment and review of V2G/G2V concepts, in conjunction with various power-electronics converter topologies. In this paper, modeling and detailed control strategies are fully designed and investigated in terms of dynamic response and harmonics. Furthermore, an extensive design and analysis of charging systems for low-duty/high-duty vehicles are also presented.

  6. Projectile fragmentation processes in 35-MeV/amu (α,xy) reactions

    International Nuclear Information System (INIS)

    Koontz, R.W.; Chang, C.C.; Holmgren, H.D.; Wu, J.R.

    1979-01-01

    Coincidence measurements with 35-MeV/amu α particles show that at least three projectile-fragmentation processes occur. The dominant process is ''absorptive'' breakup, where one component of the projectile interacts strongly with the target resulting in the emission of evaporation or nonstatistical particles while the other component behaves as a spectator. The other fragmentation processes which are observed account for only a few percent of the breakup cross section

  7. V4 AND EU INREGRATION PROCESS: THE CASE OF UKRAINE

    Directory of Open Access Journals (Sweden)

    Iryna APARINA

    2015-09-01

    Full Text Available The assistance of V4 is a very important for Ukraine. V4 countries reiterated their readiness to assist the Ukrainian authorities in preparing a comprehensive reforms package that should be at the forefront of the new government agenda. Such package should be focused on strengthening the rule of law, the efficiency of national and local governments, the transparency of public procurement, the reduction of state regulations and the fight against corruption, as well as envisaging a self-government reform that will bring the decision making process closer to the Ukrainian people. Genuinely implemented reforms, as demonstrated by V4 countries will bring the country closer to the European Union and will contribute to achieving Ukraine’s European aspirations, which are duly noted by the V4 countries. The international conference in support of Ukraine in 2015, as well as the Riga Eastern Partnership summit of 2015 represent important milestones in this respect.

  8. Computerized simulation of YAG pulse laser welding of titanium alloy (TA6V): experimental characterization and modelling of the thermomechanical aspects of this process

    International Nuclear Information System (INIS)

    Robert, Y.

    2007-09-01

    This work is a part of study which goal is to realize a computer modelling of the thermomechanical phenomena occurring during the YAG pulse laser welding of titanium alloy (TA6V). The filet welding has different heterogeneities (microstructural and mechanical). In fact, the temperature causes microstructural changes (phase transformations, precipitations) and modifies the mechanical properties. Thermomechanical modelling has thus to be established for the welding of TA6V. (author)

  9. A Design V and V Process of Hardwired PAMS for KJRR

    International Nuclear Information System (INIS)

    Jang, Gwisook; Suh, Yongsuk; Kim, Youngki

    2013-01-01

    A hardwired PAMS (Post Accident Monitoring System) of KJRR (KIJANG Research Reactor) provides the necessary information for operators to monitor and take actions after an accident condition. The hardwired PAMS are designed according to HSI (Human System Interface) guidelines. However the HSI guidelines are too high level requirements and do not reflect the detailed characteristics of commercial products. Thus, the system designers have a difficulty in the following how the system requirements are reflected in the actual and detailed requirements for purchase. Thus, a design V and V (Verification and Validation) of the hardwired PAMS should be specially managed and can be traceable through the entire development life cycle. Thus, this paper proposes a tracking matrix method for the design V and V of hardwired PAMS for KJRR. The PAMS of KJRR consists of hardwired indicators. There are many differences between the system requirements and procurement specifications for hardwired devices. The existing design V and V method for hardwired indicators is not clear. Thus, this paper proposed a tracking matrix for the design V and V of the hardwired PAMS

  10. Business process model repositories : efficient process retrieval

    NARCIS (Netherlands)

    Yan, Z.

    2012-01-01

    As organizations increasingly work in process-oriented manner, the number of business process models that they develop and have to maintain increases. As a consequence, it has become common for organizations to have collections of hundreds or even thousands of business process models. When a

  11. Modeling Aspects of Activated Sludge Processes Part l l: Mathematical Process Modeling and Biokinetics of Activated Sludge Processes

    Energy Technology Data Exchange (ETDEWEB)

    AbdElHaleem, H S [Cairo Univ.-CivlI Eng. Dept., Giza (Egypt); EI-Ahwany, A H [CairoUlmrsity- Faculty ofEngincering - Chemical Engineering Department, Giza (Egypt); Ibrahim, H I [Helwan University- Faculty of Engineering - Biomedical Engineering Department, Helwan (Egypt); Ibrahim, G [Menofia University- Faculty of Engineering Sbebin EI Kom- Basic Eng. Sc. Dept., Menofia (Egypt)

    2004-07-01

    Mathematical process modeling and biokinetics of activated sludge process were reviewed considering different types of models. It has been evaluated the task group models of ASMI. and 2, and 3 versioned by Henze et al considering the conditions of each model and the different processes of which every model consists. It is revealed that ASMI contains some defects avoided in ASM3. Relied on homogeneity, Models can be classified into homogenous models characterized by taking the activated sludge process as one phase. In this type of models, the internal mass transfer inside the floes was neglected.. Hence, the kinetic parameter produces can be considered inaccurate. The other type of models is the heterogeneous model This type considers the mass transfer operations in addition to the biochemical reaction processes; hence, the resulted kinetic parameters can be considered more accurate than that of homogenous type.

  12. Modeling Aspects of Activated Sludge Processes Part l l: Mathematical Process Modeling and Biokinetics of Activated Sludge Processes

    International Nuclear Information System (INIS)

    AbdElHaleem, H.S.; EI-Ahwany, A. H.; Ibrahim, H.I.; Ibrahim, G.

    2004-01-01

    Mathematical process modeling and biokinetics of activated sludge process were reviewed considering different types of models. It has been evaluated the task group models of ASMI. and 2, and 3 versioned by Henze et al considering the conditions of each model and the different processes of which every model consists. It is revealed that ASMI contains some defects avoided in ASM3. Relied on homogeneity, Models can be classified into homogenous models characterized by taking the activated sludge process as one phase. In this type of models, the internal mass transfer inside the floes was neglected.. Hence, the kinetic parameter produces can be considered inaccurate. The other type of models is the heterogeneous model This type considers the mass transfer operations in addition to the biochemical reaction processes; hence, the resulted kinetic parameters can be considered more accurate than that of homogenous type

  13. Process correlation analysis model for process improvement identification.

    Science.gov (United States)

    Choi, Su-jin; Kim, Dae-Kyoo; Park, Sooyong

    2014-01-01

    Software process improvement aims at improving the development process of software systems. It is initiated by process assessment identifying strengths and weaknesses and based on the findings, improvement plans are developed. In general, a process reference model (e.g., CMMI) is used throughout the process of software process improvement as the base. CMMI defines a set of process areas involved in software development and what to be carried out in process areas in terms of goals and practices. Process areas and their elements (goals and practices) are often correlated due to the iterative nature of software development process. However, in the current practice, correlations of process elements are often overlooked in the development of an improvement plan, which diminishes the efficiency of the plan. This is mainly attributed to significant efforts and the lack of required expertise. In this paper, we present a process correlation analysis model that helps identify correlations of process elements from the results of process assessment. This model is defined based on CMMI and empirical data of improvement practices. We evaluate the model using industrial data.

  14. Apoptosis imaging with Iodine-124 labeled Annexin V in Fas-mediated hepatic apoptosis model

    International Nuclear Information System (INIS)

    Lee, Tae Sup; Woo, Kwang Sun; Chung, Wee Sup; Kim, Kyung Min; Kim, Jae Hong; Chun, Kwon Soo; Choi, Chang Woon; Lim, Sang Moo; Cheon, Gi Jeong

    2006-01-01

    Healthy cells and, to a lesser extent, malignant cells undergo apoptosis or programmed cell death in response to a variety of stimuli. At an early stage in this process the cell membrane changes so that phosphatidylserine (PS), a lipid normally present on the membrane's inner surface, is exposed on the outer surface. This change in the membrane can be detected by the binding of annexin V to the external PS, and this has formed the basis for an in vitro assay for apoptosis. Blankenberg et al. have applied annexin V to the in vivo imaging of apoptosis by labeling annexin V with 99mTc. With this technique, they have been able to image apoptosis. To extend the use of annexin V to PET, it would be very desirable to iodinate the molecule. The relatively long half-life (4.2 d) of the positron emitting iodine-124 presents several advantages. For example in vivo detection and quantification of longer term biological processes is possible. Also, this cyclotron-generated radionuclide can be prepared well in advance and the established radioiodine labeling techniques can be applied. However, there are some disadvantages such as a relatively low ratio of disintegrations resulting in positrons (23%) and a rather complex decay scheme resulting in several high-energy gamma emissions (0.6- 1.69 MeV). Despite this fact, iodine-124 is still considered to be suitable for positron emission tomography (PET). In this study, we are investigating the feasibility of apoptosis imaging using iodine-124 labeled annexin V in Fas-mediated hepatic apoptosis model

  15. The impact of working memory and the "process of process modelling" on model quality: Investigating experienced versus inexperienced modellers

    DEFF Research Database (Denmark)

    Martini, Markus; Pinggera, Jakob; Neurauter, Manuel

    2016-01-01

    of reconciliation phases was positively related to PM quality in experienced modellers. Our research reveals central cognitive mechanisms in process modelling and has potential practical implications for the development of modelling software and teaching the craft of process modelling....... the role of cognitive processes as well as modelling processes in creating a PM in experienced and inexperienced modellers. Specifically, two working memory (WM) functions (holding and processing of information and relational integration) and three process of process modelling phases (comprehension...

  16. Product and Process Modelling

    DEFF Research Database (Denmark)

    Cameron, Ian T.; Gani, Rafiqul

    . These approaches are put into the context of life cycle modelling, where multiscale and multiform modelling is increasingly prevalent in the 21st century. The book commences with a discussion of modern product and process modelling theory and practice followed by a series of case studies drawn from a variety......This book covers the area of product and process modelling via a case study approach. It addresses a wide range of modelling applications with emphasis on modelling methodology and the subsequent in-depth analysis of mathematical models to gain insight via structural aspects of the models...... to biotechnology applications, food, polymer and human health application areas. The book highlights to important nature of modern product and process modelling in the decision making processes across the life cycle. As such it provides an important resource for students, researchers and industrial practitioners....

  17. 90 - GeV Higgs boson in supersymmetric models

    International Nuclear Information System (INIS)

    Grzadkowski, B.; Kalinowski, J.; Pokorski, S.

    1989-07-01

    We discuss supersymmetric models with a hierarchy of vacuum expectation values of Higgs fields. These models predict one of the physical neutral Higgs bosons to have its mass very close to the Z-boson mass. Properties of such a 90-GeV Higgs boson are discussed. (author)

  18. Modeling of biopharmaceutical processes. Part 2: Process chromatography unit operation

    DEFF Research Database (Denmark)

    Kaltenbrunner, Oliver; McCue, Justin; Engel, Philip

    2008-01-01

    Process modeling can be a useful tool to aid in process development, process optimization, and process scale-up. When modeling a chromatography process, one must first select the appropriate models that describe the mass transfer and adsorption that occurs within the porous adsorbent. The theoret......Process modeling can be a useful tool to aid in process development, process optimization, and process scale-up. When modeling a chromatography process, one must first select the appropriate models that describe the mass transfer and adsorption that occurs within the porous adsorbent...

  19. Experimental Constraints on γ-Ray Pulsar Gap Models and the Pulsar GeV to Pulsar Wind Nebula TeV Connection

    Science.gov (United States)

    Abeysekara, A. U.; Linnemann, J. T.

    2015-05-01

    The pulsar emission mechanism in the gamma ray energy band is poorly understood. Currently, there are several models under discussion in the pulsar community. These models can be constrained by studying the collective properties of a sample of pulsars, which became possible with the large sample of gamma ray pulsars discovered by the Fermi Large Area Telescope. In this paper we develop a new experimental multi-wavelength technique to determine the beaming factor ≤ft( {{f}{Ω }} \\right) dependance on spin-down luminosity of a set of GeV pulsars. This technique requires three input parameters: pulsar spin-down luminosity, pulsar phase-averaged GeV flux, and TeV or X-ray flux from the associated pulsar wind nebula (PWN). The analysis presented in this paper uses the PWN TeV flux measurements to study the correlation between {{f}{Ω }} and \\dot{E}. The measured correlation has some features that favor the Outer Gap model over the Polar Cap, Slot Gap, and One Pole Caustic models for pulsar emission in the energy range of 0.1-100 GeV, but one must keep in mind that these simulated models failed to explain many of the most important pulsar population characteristics. A tight correlation between the pulsar GeV emission and PWN TeV emission was also observed, which suggests the possibility of a linear relationship between the two emission mechanisms. In this paper we also discuss a possible mechanism to explain this correlation.

  20. The impact of working memory and the "process of process modelling" on model quality: Investigating experienced versus inexperienced modellers.

    Science.gov (United States)

    Martini, Markus; Pinggera, Jakob; Neurauter, Manuel; Sachse, Pierre; Furtner, Marco R; Weber, Barbara

    2016-05-09

    A process model (PM) represents the graphical depiction of a business process, for instance, the entire process from online ordering a book until the parcel is delivered to the customer. Knowledge about relevant factors for creating PMs of high quality is lacking. The present study investigated the role of cognitive processes as well as modelling processes in creating a PM in experienced and inexperienced modellers. Specifically, two working memory (WM) functions (holding and processing of information and relational integration) and three process of process modelling phases (comprehension, modelling, and reconciliation) were related to PM quality. Our results show that the WM function of relational integration was positively related to PM quality in both modelling groups. The ratio of comprehension phases was negatively related to PM quality in inexperienced modellers and the ratio of reconciliation phases was positively related to PM quality in experienced modellers. Our research reveals central cognitive mechanisms in process modelling and has potential practical implications for the development of modelling software and teaching the craft of process modelling.

  1. Applying the Business Process and Practice Alignment Meta-model: Daily Practices and Process Modelling

    Directory of Open Access Journals (Sweden)

    Ventura Martins Paula

    2017-03-01

    Full Text Available Background: Business Process Modelling (BPM is one of the most important phases of information system design. Business Process (BP meta-models allow capturing informational and behavioural aspects of business processes. Unfortunately, standard BP meta-modelling approaches focus just on process description, providing different BP models. It is not possible to compare and identify related daily practices in order to improve BP models. This lack of information implies that further research in BP meta-models is needed to reflect the evolution/change in BP. Considering this limitation, this paper introduces a new BP meta-model designed by Business Process and Practice Alignment Meta-model (BPPAMeta-model. Our intention is to present a meta-model that addresses features related to the alignment between daily work practices and BP descriptions. Objectives: This paper intends to present a metamodel which is going to integrate daily work information into coherent and sound process definitions. Methods/Approach: The methodology employed in the research follows a design-science approach. Results: The results of the case study are related to the application of the proposed meta-model to align the specification of a BP model with work practices models. Conclusions: This meta-model can be used within the BPPAM methodology to specify or improve business processes models based on work practice descriptions.

  2. Study of two photon production process in proton-proton collisions at 216 MeV

    International Nuclear Information System (INIS)

    Khrykin, A.S.

    2002-01-01

    The energy spectrum for high energy γ-rays (Eγ ≥ 10 MeV) from the process pp → γγX emitted at 90 deg. in the laboratory frame has been measured at 216 MeV. The resulting photon energy spectrum extracted from γ - γ coincidence events consists of a narrow peak (5.3σ) at a photon energy of about 24 MeV and a relatively broad peak (3.5σ) in the energy range of (50 - 70) MeV. This behavior of the photon energy spectrum is interpreted as a signature of the exotic dibaryon resonance d 1 * with a mass of about 1956 MeV which is assumed to be formed in the radiative process pp → γd 1 * followed by its electromagnetic decay via the d 1 * → ppγ mode. The experimental spectrum is compared with those obtained by means of Monte Carlo simulations

  3. Modeling and simulation of plasma materials processing devices

    International Nuclear Information System (INIS)

    Graves, D.B.

    1996-01-01

    Plasma processing has emerged as a central technology in the manufacture of integrated circuits (ICs) and related industries. These plasmas are weakly to partially ionized gases, typically operated at a few to several hundred mTorr gas pressure, with neutral temperatures ranging from room temperature to 500 degrees K. Electron mean energies are typically a few eV and ion energies in the bulk plasma are about 0.05-0.5 eV. Positive ions axe accelerated in the sheaths to impact surfaces with energies ranging from about 10 eV to hundreds of eV. These energetic ions profoundly affect rates of surface chemical reactions. One of the consequences of the recent rapid growth in the IC industry has been a greater focus on manufacturing productivity. The capital costs of equipment that is used in manufacturing IC's has become a large fraction of the ∼ $1 billion cost of building a wafer fab. There is now a strong economic incentive to develop workstation-based simulations of plasma chemical reactors in order to design, optimize and control plasma reactors. I will summarize efforts to develop such models, including electromagnetic coupling, and transport and kinetics of charged and neutral species. Length and time scale disparities in the plasma tool challenge current simulation approaches, and I will address strategies to attack aspects of this problem. In addition, I will present some of our recent efforts to exploit molecular dynamics simulations employing empirical potentials to get hints about qualitative mechanisms and ideas on how to formulate rate expressions for plasma-surface chemical processes. Video illustrations of selected sets of ion trajectories impacting near-surface regions of the substrate will be presented

  4. Recovering probabilities for nucleotide trimming processes for T cell receptor TRA and TRG V-J junctions analyzed with IMGT tools

    Directory of Open Access Journals (Sweden)

    Lefranc Marie-Paule

    2008-10-01

    Full Text Available Abstract Background Nucleotides are trimmed from the ends of variable (V, diversity (D and joining (J genes during immunoglobulin (IG and T cell receptor (TR rearrangements in B cells and T cells of the immune system. This trimming is followed by addition of nucleotides at random, forming the N regions (N for nucleotides of the V-J and V-D-J junctions. These processes are crucial for creating diversity in the immune response since the number of trimmed nucleotides and the number of added nucleotides vary in each B or T cell. IMGT® sequence analysis tools, IMGT/V-QUEST and IMGT/JunctionAnalysis, are able to provide detailed and accurate analysis of the final observed junction nucleotide sequences (tool "output". However, as trimmed nucleotides can potentially be replaced by identical N region nucleotides during the process, the observed "output" represents a biased estimate of the "true trimming process." Results A probabilistic approach based on an analysis of the standardized tool "output" is proposed to infer the probability distribution of the "true trimmming process" and to provide plausible biological hypotheses explaining this process. We collated a benchmark dataset of TR alpha (TRA and TR gamma (TRG V-J rearranged sequences and junctions analysed with IMGT/V-QUEST and IMGT/JunctionAnalysis, the nucleotide sequence analysis tools from IMGT®, the international ImMunoGeneTics information system®, http://imgt.cines.fr. The standardized description of the tool output is based on the IMGT-ONTOLOGY axioms and concepts. We propose a simple first-order model that attempts to transform the observed "output" probability distribution into an estimate closer to the "true trimming process" probability distribution. We use this estimate to test the hypothesis that Poisson processes are involved in trimming. This hypothesis was not rejected at standard confidence levels for three of the four trimming processes: TRAV, TRAJ and TRGV. Conclusion By

  5. Recovering probabilities for nucleotide trimming processes for T cell receptor TRA and TRG V-J junctions analyzed with IMGT tools.

    Science.gov (United States)

    Bleakley, Kevin; Lefranc, Marie-Paule; Biau, Gérard

    2008-10-02

    Nucleotides are trimmed from the ends of variable (V), diversity (D) and joining (J) genes during immunoglobulin (IG) and T cell receptor (TR) rearrangements in B cells and T cells of the immune system. This trimming is followed by addition of nucleotides at random, forming the N regions (N for nucleotides) of the V-J and V-D-J junctions. These processes are crucial for creating diversity in the immune response since the number of trimmed nucleotides and the number of added nucleotides vary in each B or T cell. IMGT sequence analysis tools, IMGT/V-QUEST and IMGT/JunctionAnalysis, are able to provide detailed and accurate analysis of the final observed junction nucleotide sequences (tool "output"). However, as trimmed nucleotides can potentially be replaced by identical N region nucleotides during the process, the observed "output" represents a biased estimate of the "true trimming process." A probabilistic approach based on an analysis of the standardized tool "output" is proposed to infer the probability distribution of the "true trimmming process" and to provide plausible biological hypotheses explaining this process. We collated a benchmark dataset of TR alpha (TRA) and TR gamma (TRG) V-J rearranged sequences and junctions analysed with IMGT/V-QUEST and IMGT/JunctionAnalysis, the nucleotide sequence analysis tools from IMGT, the international ImMunoGeneTics information system, http://imgt.cines.fr. The standardized description of the tool output is based on the IMGT-ONTOLOGY axioms and concepts. We propose a simple first-order model that attempts to transform the observed "output" probability distribution into an estimate closer to the "true trimming process" probability distribution. We use this estimate to test the hypothesis that Poisson processes are involved in trimming. This hypothesis was not rejected at standard confidence levels for three of the four trimming processes: TRAV, TRAJ and TRGV. By using trimming of rearranged TR genes as a benchmark, we

  6. Modeling multiphase materials processes

    CERN Document Server

    Iguchi, Manabu

    2010-01-01

    ""Modeling Multiphase Materials Processes: Gas-Liquid Systems"" describes the methodology and application of physical and mathematical modeling to multi-phase flow phenomena in materials processing. The book focuses on systems involving gas-liquid interaction, the most prevalent in current metallurgical processes. The performance characteristics of these processes are largely dependent on transport phenomena. This volume covers the inherent characteristics that complicate the modeling of transport phenomena in such systems, including complex multiphase structure, intense turbulence, opacity of

  7. The M/V Selendang Ayu commercial fisheries work group process

    International Nuclear Information System (INIS)

    DeCola, E.; Robertson, T.; Pearson, L.; Dietrick, L.; Folley, G.

    2006-01-01

    The process used to form and operate the Selendang Ayu Fisheries Work Group was discussed. The M/V Selendang Ayu ran aground and broke apart near Unalaska Island, Alaska in December 2004, causing immediate concern to nearby commercial fisheries and seafood processors. The work group, which was formed less than 2 weeks after the initial spill, was modeled after other Alaska work groups and consisted of experts in environmental health, fisheries biology and oceanography. Its task was to facilitate communication between fishery managers and spill responders to identify risk-reducing measures and to minimize impacts to fisheries in Alaska. In support of Alaska's zero tolerance policy toward any contamination of food processed in Alaska, the working group conducted water quality sampling in and around Unalaska Island from December 2004 though March 2005. Information was gathered regarding fishery opening and closing dates, geographic extent of fisheries, gear and vessel types, seawater circulation systems, fish processing operations, vessel transit routes and other considerations related to the potential exposure of vessels, gear or seafood to oil. This paper described the different pathways that commercial fish species may be exposed to spilled oil, along with the major fish species at risk. One small fishery closure was enacted. The working group was successful in ensuring that major commercial fisheries for snow crab, pollock, halibut and other species were successfully conducted during the spill, with all loads undergoing mandatory seafood inspections. Seafood products were not contaminated and market prices remained stable. 15 refs., 6 tabs., 4 figs

  8. A model for Simpson's 17 keV neutrino

    International Nuclear Information System (INIS)

    Rajpoot, S.

    1992-01-01

    Recent studies of β-decay spectra seem to confirm Simpson's earlier findings that the electron neutrinos contain a small (1%) admixture of a 17 keV Dirac neutrino. In this paper an unconventional model with SU(2) L x SU(2) R x U(1) B-1 gauge interactions is presented in which all neutrinos are Dirac particles. Electron and muon neutrinos acquire seesaw Dirac masses of order 10 -3 eV for the MSW solution for the solar neutrino problem. The τ neutrino is identified as Simpson's 17 keV neutrino. Constraints coming from cosmology and particle physics are shown to be satisfied

  9. An extended car-following model at un-signalized intersections under V2V communication environment

    Science.gov (United States)

    Wang, Tao; Li, Peng

    2018-01-01

    An extended car-following model is proposed in this paper to analyze the impacts of V2V (vehicle to vehicle) communication on the micro driving behavior at the un-signalized intersection. A four-leg un-signalized intersection with twelve streams (left-turn, through movement, and right turn from each leg) is used. The effect of the guidance strategy on the reduction of the rate of stops and total delay is explored by comparing the proposed model and the traditional FVD car-following model. The numerical results illustrate that potential conflicts between vehicles can be predicted and some stops can be avoided by decelerating in advance. The driving comfort and traffic efficiency can be improved accordingly. More benefits could be obtained under the long communication range, low to medium traffic density, and simple traffic pattern conditions. PMID:29425243

  10. PROCESS MODELLING AND DEBOTTLENECKING STUDY OF A VACCINE PRODUCTION

    Directory of Open Access Journals (Sweden)

    Nurul Huda Mohamed Safri

    2012-04-01

    Full Text Available ABSTRACT: The main objective of this research work was to model and optimise the production of a locally-developed Infectious Coryza (IC vaccine. The simulation work was performed using a commercially available batch process simulator SuperPro Designer v5.5. Six debottlenecking schemes were analysed using throughput analysis and cost to benefit ratio (CBR when the annual production was set to increase by 100%. Based on the economic analysis, the selected debottlenecking scheme has an annual predicted revenue of USD 240 million, with a gross margin of 9.13% and a return on investment (ROI of 46.12%. In addition, the payback period of the selected scheme is estimated to be within three years. ABSTRAK: Objektif utama dalam penyelidikan ini adalah untuk memodelkan dan mengoptimumkan hasil pembuatan vaksin tempatan Coryza berjangkit. Kerja simulasi ini dijalankan menggunakan alat simulasi Super Pro Designer v5.5. Sebanyak enam (6 skema khusus diujikaji menggunakan analisis pemprosesan dan kos kepada nisbah faedah (CBR apabila pembuatan tahunan meningkat kepada 100%. Berdasarkan analisis ekonomi yang telah dilakukan, sesuatu skema khusus yang dipilih mempunyai keuntungan sebanyak USD 240 juta dengan margin kasar 9.13% dan pulangan atas pelaburan (ROI sebanyak 46.12%. Selain itu juga, tempoh pembayaran balik bagi skema yang dipilih dianggarkan dalam tempoh tiga(3 tahun.KEYWORDS: process simulation; modelling; debottlenecking; optimisation

  11. GeoSciML v3.0 - a significant upgrade of the CGI-IUGS geoscience data model

    Science.gov (United States)

    Raymond, O.; Duclaux, G.; Boisvert, E.; Cipolloni, C.; Cox, S.; Laxton, J.; Letourneau, F.; Richard, S.; Ritchie, A.; Sen, M.; Serrano, J.-J.; Simons, B.; Vuollo, J.

    2012-04-01

    analytical data using the Observations and Measurements (ISO19156) and SWE Common v2 models. The GeoSciML v3 data model does not include vocabularies to support the data model. However, it does provide a standard pattern to reference controlled vocabulary concepts using HTTP-URIs. The international GeoSciML community has developed distributed RDF-based geoscience vocabularies that can be accessed by GeoSciML web services using the standard pattern recommended in GeoSciML v3. GeoSciML v3 is the first version of GeoSciML that will be accompanied by web service validation tools using Schematron rules. For example, these validation tools may check for compliance of a web service to a particular profile of GeoSciML, or for logical consistency of data content that cannot be enforced by the application schemas. This validation process will support accreditation of GeoSciML services and a higher degree of semantic interoperability. * International Union of Geological Sciences Commission for Management and Application of Geoscience Information (CGI-IUGS)

  12. Semantic Business Process Modeling

    OpenAIRE

    Markovic, Ivan

    2010-01-01

    This book presents a process-oriented business modeling framework based on semantic technologies. The framework consists of modeling languages, methods, and tools that allow for semantic modeling of business motivation, business policies and rules, and business processes. Quality of the proposed modeling framework is evaluated based on the modeling content of SAP Solution Composer and several real-world business scenarios.

  13. Business Process Modeling: Perceived Benefits

    Science.gov (United States)

    Indulska, Marta; Green, Peter; Recker, Jan; Rosemann, Michael

    The process-centered design of organizations and information systems is globally seen as an appropriate response to the increased economic pressure on organizations. At the methodological core of process-centered management is process modeling. However, business process modeling in large initiatives can be a time-consuming and costly exercise, making it potentially difficult to convince executive management of its benefits. To date, and despite substantial interest and research in the area of process modeling, the understanding of the actual benefits of process modeling in academia and practice is limited. To address this gap, this paper explores the perception of benefits derived from process modeling initiatives, as reported through a global Delphi study. The study incorporates the views of three groups of stakeholders - academics, practitioners and vendors. Our findings lead to the first identification and ranking of 19 unique benefits associated with process modeling. The study in particular found that process modeling benefits vary significantly between practitioners and academics. We argue that the variations may point to a disconnect between research projects and practical demands.

  14. GATE V6: a major enhancement of the GATE simulation platform enabling modelling of CT and radiotherapy

    Energy Technology Data Exchange (ETDEWEB)

    Jan, S; Becheva, E [DSV/I2BM/SHFJ, Commissariat a l' Energie Atomique, Orsay (France); Benoit, D; Rehfeld, N; Stute, S; Buvat, I [IMNC-UMR 8165 CNRS-Paris 7 and Paris 11 Universities, 15 rue Georges Clemenceau, 91406 Orsay Cedex (France); Carlier, T [INSERM U892-Cancer Research Center, University of Nantes, Nantes (France); Cassol, F; Morel, C [Centre de physique des particules de Marseille, CNRS-IN2P3 and Universite de la Mediterranee, Aix-Marseille II, 163, avenue de Luminy, 13288 Marseille Cedex 09 (France); Descourt, P; Visvikis, D [INSERM, U650, Laboratoire du Traitement de l' Information Medicale (LaTIM), CHU Morvan, Brest (France); Frisson, T; Grevillot, L; Guigues, L; Sarrut, D; Zahra, N [Universite de Lyon, CREATIS, CNRS UMR5220, Inserm U630, INSA-Lyon, Universite Lyon 1, Centre Leon Berard (France); Maigne, L; Perrot, Y [Laboratoire de Physique Corpusculaire, 24 Avenue des Landais, 63177 Aubiere Cedex (France); Schaart, D R [Delft University of Technology, Radiation Detection and Medical Imaging, Mekelweg 15, 2629 JB Delft (Netherlands); Pietrzyk, U, E-mail: buvat@imnc.in2p3.fr [Reseach Center Juelich, Institute of Neurosciences and Medicine and Department of Physics, University of Wuppertal (Germany)

    2011-02-21

    GATE (Geant4 Application for Emission Tomography) is a Monte Carlo simulation platform developed by the OpenGATE collaboration since 2001 and first publicly released in 2004. Dedicated to the modelling of planar scintigraphy, single photon emission computed tomography (SPECT) and positron emission tomography (PET) acquisitions, this platform is widely used to assist PET and SPECT research. A recent extension of this platform, released by the OpenGATE collaboration as GATE V6, now also enables modelling of x-ray computed tomography and radiation therapy experiments. This paper presents an overview of the main additions and improvements implemented in GATE since the publication of the initial GATE paper (Jan et al 2004 Phys. Med. Biol. 49 4543-61). This includes new models available in GATE to simulate optical and hadronic processes, novelties in modelling tracer, organ or detector motion, new options for speeding up GATE simulations, examples illustrating the use of GATE V6 in radiotherapy applications and CT simulations, and preliminary results regarding the validation of GATE V6 for radiation therapy applications. Upon completion of extensive validation studies, GATE is expected to become a valuable tool for simulations involving both radiotherapy and imaging.

  15. Stochastic series expansion simulation of the t -V model

    Science.gov (United States)

    Wang, Lei; Liu, Ye-Hua; Troyer, Matthias

    2016-04-01

    We present an algorithm for the efficient simulation of the half-filled spinless t -V model on bipartite lattices, which combines the stochastic series expansion method with determinantal quantum Monte Carlo techniques widely used in fermionic simulations. The algorithm scales linearly in the inverse temperature, cubically with the system size, and is free from the time-discretization error. We use it to map out the finite-temperature phase diagram of the spinless t -V model on the honeycomb lattice and observe a suppression of the critical temperature of the charge-density-wave phase in the vicinity of a fermionic quantum critical point.

  16. Cloning, preparation and preliminary crystallographic studies of penicillin V acylase autoproteolytic processing mutants

    International Nuclear Information System (INIS)

    Chandra, P. Manish; Brannigan, James A.; Prabhune, Asmita; Pundle, Archana; Turkenburg, Johan P.; Dodson, G. Guy; Suresh, C. G.

    2004-01-01

    The production, crystallization and characterization of three inactive mutants of penicillin V acylase from B. sphaericus in their respective precursor and processed forms are reported. The space groups are different for the native enzyme and the mutants. The crystallization of three catalytically inactive mutants of penicillin V acylase (PVA) from Bacillus sphaericus in precursor and processed forms is reported. The mutant proteins crystallize in different primitive monoclinic space groups that are distinct from the crystal forms for the native enzyme. Directed mutants and clone constructs were designed to study the post-translational autoproteolytic processing of PVA. The catalytically inactive mutants will provide three-dimensional structures of precursor PVA forms, plus open a route to the study of enzyme–substrate complexes for this industrially important enzyme

  17. Modeling of air-gap membrane distillation process: A theoretical and experimental study

    KAUST Repository

    Alsaadi, Ahmad Salem

    2013-06-03

    A one dimensional (1-D) air gap membrane distillation (AGMD) model for flat sheet type modules has been developed. This model is based on mathematical equations that describe the heat and mass transfer mechanisms of a single-stage AGMD process. It can simulate AGMD modules in both co-current and counter-current flow regimes. The theoretical model was validated using AGMD experimental data obtained under different operating conditions and parameters. The predicted water vapor flux was compared to the flux measured at five different feed water temperatures, two different feed water salinities, three different air gap widths and two MD membranes with different average pore sizes. This comparison showed that the model flux predictions are strongly correlated with the experimental data, with model predictions being within +10% of the experimentally determined values. The model was then used to study and analyze the parameters that have significant effect on scaling-up the AGMD process such as the effect of increasing the membrane length, and feed and coolant flow rates. The model was also used to analyze the maximum thermal efficiency of the AGMD process by tracing changes in water production rate and the heat input to the process along the membrane length. This was used to understand the gain in both process production and thermal efficiency for different membrane surface areas and the resultant increases in process capital and water unit cost. © 2013 Elsevier B.V.

  18. Investigation on mechanical alloying process for v-cr-ti alloys

    International Nuclear Information System (INIS)

    Stanciulescu, M.; Carlan, P.; Mihalache, M.; Bucsa, G.; Abrudeanu, M.; Galateanu, A.

    2015-01-01

    Mechanical alloying (MA) is an efficient approach for fabricating oxide-dispersion alloys and structural materials including vanadium alloys for fusion and fission application. Dissolution behaviour of the alloying elements is a key issue for optimizing the mechanical alloying process in fabricating vanadium alloys. This paper studies the MA process of V-4wt.%Cr-4wt.%Ti alloy. The outcomes of the MA powders in a planetary ball mill are reported in terms of powder particle size and morphology evolution and elemental composition. The impact of spark-plasma sintering process on the mechanically alloyed powder is analysed. An optimal set of sintering parameters, including the maximum temperature, the dwell time and the heating rate are determined. (authors)

  19. Intrinsic Nonlinearities and Layout Impacts of 100 V Integrated Power MOSFETs in Partial SOI Process

    DEFF Research Database (Denmark)

    Fan, Lin; Knott, Arnold; Jørgensen, Ivan Harald Holger

    Parasitic capacitances of power semiconductors are a part of the key design parameters of state-of-the-art very high frequency (VHF) power supplies. In this poster, four 100 V integrated power MOSFETs with different layout structures are designed, implemented, and analyzed in a 0.18 ȝm partial...... Silicon-on-Insulator (SOI) process with a die area 2.31 mm2.  A small-signal model of power MOSFETs is proposed to systematically analyze the nonlinear parasitic capacitances in different transistor states: off-state, sub-threshold region, and on-state in the linear region. 3D plots are used to summarize...

  20. Transferable tight binding model for strained group IV and III-V heterostructures

    Science.gov (United States)

    Tan, Yaohua; Povolotskyi, Micheal; Kubis, Tillmann; Boykin, Timothy; Klimeck, Gerhard

    Modern semiconductor devices have reached critical device dimensions in the range of several nanometers. For reliable prediction of device performance, it is critical to have a numerical efficient model that are transferable to material interfaces. In this work, we present an empirical tight binding (ETB) model with transferable parameters for strained IV and III-V group semiconductors. The ETB model is numerically highly efficient as it make use of an orthogonal sp3d5s* basis set with nearest neighbor inter-atomic interactions. The ETB parameters are generated from HSE06 hybrid functional calculations. Band structures of strained group IV and III-V materials by ETB model are in good agreement with corresponding HSE06 calculations. Furthermore, the ETB model is applied to strained superlattices which consist of group IV and III-V elements. The ETB model turns out to be transferable to nano-scale hetero-structure. The ETB band structures agree with the corresponding HSE06 results in the whole Brillouin zone. The ETB band gaps of superlattices with common cations or common anions have discrepancies within 0.05eV.

  1. Growth of α-V{sub 2}O{sub 5} nanostructured thin films as a function of deposition process

    Energy Technology Data Exchange (ETDEWEB)

    Singh, Megha, E-mail: meghasingh.life@gmail.com; Sharma, Rabindar K.; Reddy, G. B. [Thin film Laboratory, Department of Physics, Indian Institute of Technology Delhi -110016 (India)

    2016-05-23

    In this communication, we synthesizedvanadium pentoxide (α-V{sub 2}O{sub 5}) nanostructured thin films (NSTs) using four different methods for obtaining vanadate species namely thermal evaporation (source of vanadate species are V{sub 2}O{sub 5} powder and vanadium metal foil) and plasma assisted sublimation process (source of vanadate species are V{sub 2}O{sub 5} powder and vanadium metal foil). The effect of plasmaon morphological and structural propertieshave been systematicallystudied. XRD revealed thermal evaporation process yielded amorphous films whereassublimation process yielded highly crystalline α-V{sub 2}O{sub 5} films. HRTEM of nanobelts show, the growth is preferred in (001) crystallographic direction with interplanar distance of 0.43 nm. XPS revealed O/V ratio of ~2.4, which nearly agrees with standard V{sub 2}O{sub 5} stoichiometry. SEM revealed deposition process affect morphology of films; thermal evaporation results in smoother film while plasma assisted sublimation process reveals nanoflakes and nanobelts (NBs). All the results are inconcordance with each other.

  2. Adenoviral protein V promotes a process of viral assembly through nucleophosmin 1

    Energy Technology Data Exchange (ETDEWEB)

    Ugai, Hideyo; Dobbins, George C.; Wang, Minghui [Division of Human Gene Therapy, Departments of Medicine, Obstetrics and Gynecology, Pathology, and Surgery, University of Alabama at Birmingham, Birmingham, AL 35294 (United States); Le, Long P. [Massachusetts General Hospital, Pathology Service, 55 Fruit St.-GRJ 249, Boston, MA 02114 (United States); Matthews, David A. [School of Cellular and Molecular Medicine, Medical Sciences Building, University of Bristol, Bristol BS8 1TD (United Kingdom); Curiel, David T., E-mail: dcuriel@radonc.wustl.edu [Division of Human Gene Therapy, Departments of Medicine, Obstetrics and Gynecology, Pathology, and Surgery, University of Alabama at Birmingham, Birmingham, AL 35294 (United States); The Gene Therapy Center, University of Alabama at Birmingham, Birmingham, AL 35294 (United States)

    2012-10-25

    Adenoviral infection induces nucleoplasmic redistribution of a nucleolar nucleophosmin 1/NPM1/B23.1. NPM1 is preferentially localized in the nucleoli of normal cells, whereas it is also present at the nuclear matrix in cancer cells. However, the biological roles of NPM1 during infection are unknown. Here, by analyzing a pV-deletion mutant, Ad5-dV/TSB, we demonstrate that pV promotes the NPM1 translocation from the nucleoli to the nucleoplasm in normal cells, and the NPM1 translocation is correlated with adenoviral replication. Lack of pV causes a dramatic reduction of adenoviral replication in normal cells, but not cancer cells, and Ad5-dV/TSB was defective in viral assembly in normal cells. NPM1 knockdown inhibits adenoviral replication, suggesting an involvement of NPM1 in adenoviral biology. Further, we show that NPM1 interacts with empty adenovirus particles which are an intermediate during virion maturation by immunoelectron microscopy. Collectively, these data implicate that pV participates in a process of viral assembly through NPM1.

  3. Quantum state-to-state dynamics for the quenching process of Br(2P1/2) + H2(v(i) = 0, 1, j(i) = 0).

    Science.gov (United States)

    Xie, Changjian; Jiang, Bin; Xie, Daiqian; Sun, Zhigang

    2012-03-21

    Quantum state-to-state dynamics for the quenching process Br((2)P(1/2)) + H(2)(v(i) = 0, 1, j(i) = 0) → Br((2)P(3/2)) + H(2)(v(f), j(f)) has been studied based on two-state model on the recent coupled potential energy surfaces. It was found that the quenching probabilities have some oscillatory structures due to the interference of reflected flux in the Br((2)P(1/2)) + H(2) and Br((2)P(3/2)) + H(2) channels by repulsive potential in the near-resonant electronic-to-vibrational energy transfer process. The final vibrational state resolved integral cross sections were found to be dominated by the quenching process Br((2)P(1/2)) + H(2)(v) → Br((2)P(3/2)) + H(2)(v+1) and the nonadiabatic reaction probabilities for Br((2)P(1/2)) + H(2)(v = 0, 1, j(i) = 0) are quite small, which are consistent with previous theoretical and experimental results. Our calculated total quenching rate constant for Br((2)P(1/2)) + H(2)(v(i) = 0, j(i) = 0) at room temperature is in good agreement with the available experimental data. © 2012 American Institute of Physics

  4. Scalability of the LEU-Modified Cintichem Process: 3-MeV Van de Graaff and 35-MeV Electron Linear Accelerator Studies

    Energy Technology Data Exchange (ETDEWEB)

    Rotsch, David A. [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Brossard, Tom [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Roussin, Ethan [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Quigley, Kevin [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Chemerisov, Sergey [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Gromov, Roman [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Jonah, Charles [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Hafenrichter, Lohman [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Tkac, Peter [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Krebs, John [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division; Vandegrift, George F. [Argonne National Lab. (ANL), Argonne, IL (United States). Nuclear Engineering Division

    2016-10-31

    Molybdenum-99, the mother of Tc-99m, can be produced from fission of U-235 in nuclear reactors and purified from fission products by the Cintichem process, later modified for low-enriched uranium (LEU) targets. The key step in this process is the precipitation of Mo with α-benzoin oxime (ABO). The stability of this complex to radiation has been examined. Molybdenum-ABO was irradiated with 3 MeV electrons produced by a Van de Graaff generator and 35 MeV electrons produced by a 50 MeV/25 kW electron linear accelerator. Dose equivalents of 1.7–31.2 kCi of Mo-99 were administered to freshly prepared Mo-ABO. Irradiated samples of Mo-ABO were processed according to the LEU Modified-Cintichem process. The Van de Graaff data indicated good radiation stability of the Mo-ABO complex up to ~15 kCi dose equivalents of Mo-99 and nearly complete destruction at doses >24 kCi Mo-99. The linear accelerator data indicate that even at 6.2 kCi of Mo-99 equivalence of dose, the sample lost ~20% of Mo-99. The 20% loss of Mo-99 at this low dose may be attributed to thermal decomposition of the product from the heat deposited in the sample during irradiation.

  5. Breast Malignant Processes Echography at V. I. Lenin Hospital

    International Nuclear Information System (INIS)

    Tamayo Hernandez, Lissette; Aguilar Munnoz, Liliana; Estevez de la Torre, Elyanne; Ricardo Guerrero, Migdania; Rodriguez Infanzon, Olga; Hung Arroyo, Jose Manuel

    2007-01-01

    A descriptive and cross-sectional study was carried out in patients with a presumptive diagnosis of breast malignant processes . The patients came from the breast cancer consulting room at V. I. Lenin Hospital from January 2005 to December 2005. A breast ultrasound imaging was done aimed at deepening in the diagnosis which was predominant in pre- menopausal women , being the left breast external upper quadrant the most frequent site of localization. The most suggestive echography pattern of malignant processes was the image with changes in its echostructure, soft echos in its inner part and irregular edge were observed. According to the mammography results the most frequent presentation was the visualization of a thick breast of irregular edge with spicule formations. There was a diagnostic correspondence between the echogram and the mammography considering that this is an adequate method for the initial evaluation of breast malignant processes

  6. Preliminary Evaluation on Characteristics of Waste from Pyro-processing (FS v5.1)

    International Nuclear Information System (INIS)

    Kim, In-Young; Choi, Heui-Joo

    2016-01-01

    In this study, characteristic of waste from pyro-processing which based on material balance FS v5.1 is evaluated for revising of A-KRS concept. To reduce volume and toxicity of PWR SNFs, the P and T Technology using pyro-processing and SFR is under development in KAERI. In accordance with this R and D, the A-KRS was developed by KAERI for disposal of waste from pyro-processing. After A-KRS concept development, material balance has been revised and characteristics of waste have been changed. To solve impending saturation of storage capacity of NPP sites, national policy on SNF management seems to be determined shortly. Demand on detailed analysis on impact of P and T using pyro-processing and SFR as base data is increasing. To compare direct disposal scenario to the P and T scenario using pyro-processing and SFR, updating of A-KRS reflecting amendment of material balance is required. In this study, characteristic of waste from pyro-processing which based on material balance FS v5.1 is evaluated. Every 15 types of outputs generated from pyro-processing are evaluated. Great reduction can be achieved by complete reuse of U/TRU/RE in SFR, because most of decay heat, radioactivity and radiotoxicity are generated from U/TRU/RE ingot. Within about 300 years, the fly ash filter containing Cs, Sr containing salt waste are important waste in perspective of decay heat and radioactivity. Importance of RE containing salt waste reduced. Management of hold-ups must be clarified because hold-ups take large portion of decay heat, radioactivity in the material balance FS v5.1

  7. Preliminary Evaluation on Characteristics of Waste from Pyro-processing (FS v5.1)

    Energy Technology Data Exchange (ETDEWEB)

    Kim, In-Young; Choi, Heui-Joo [Korea Atomic Energy Research Institute, Daejeon (Korea, Republic of)

    2016-10-15

    In this study, characteristic of waste from pyro-processing which based on material balance FS v5.1 is evaluated for revising of A-KRS concept. To reduce volume and toxicity of PWR SNFs, the P and T Technology using pyro-processing and SFR is under development in KAERI. In accordance with this R and D, the A-KRS was developed by KAERI for disposal of waste from pyro-processing. After A-KRS concept development, material balance has been revised and characteristics of waste have been changed. To solve impending saturation of storage capacity of NPP sites, national policy on SNF management seems to be determined shortly. Demand on detailed analysis on impact of P and T using pyro-processing and SFR as base data is increasing. To compare direct disposal scenario to the P and T scenario using pyro-processing and SFR, updating of A-KRS reflecting amendment of material balance is required. In this study, characteristic of waste from pyro-processing which based on material balance FS v5.1 is evaluated. Every 15 types of outputs generated from pyro-processing are evaluated. Great reduction can be achieved by complete reuse of U/TRU/RE in SFR, because most of decay heat, radioactivity and radiotoxicity are generated from U/TRU/RE ingot. Within about 300 years, the fly ash filter containing Cs, Sr containing salt waste are important waste in perspective of decay heat and radioactivity. Importance of RE containing salt waste reduced. Management of hold-ups must be clarified because hold-ups take large portion of decay heat, radioactivity in the material balance FS v5.1.

  8. WWTP Process Tank Modelling

    DEFF Research Database (Denmark)

    Laursen, Jesper

    The present thesis considers numerical modeling of activated sludge tanks on municipal wastewater treatment plants. Focus is aimed at integrated modeling where the detailed microbiological model the Activated Sludge Model 3 (ASM3) is combined with a detailed hydrodynamic model based on a numerical...... solution of the Navier-Stokes equations in a multiphase scheme. After a general introduction to the activated sludge tank as a system, the activated sludge tank model is gradually setup in separate stages. The individual sub-processes that are often occurring in activated sludge tanks are initially...... hydrofoil shaped propellers. These two sub-processes deliver the main part of the supplied energy to the activated sludge tank, and for this reason they are important for the mixing conditions in the tank. For other important processes occurring in the activated sludge tank, existing models and measurements...

  9. The impact of global nuclear mass model uncertainties on r-process abundance predictions

    Directory of Open Access Journals (Sweden)

    Mumpower M.

    2015-01-01

    Full Text Available Rapid neutron capture or ‘r-process’ nucleosynthesis may be responsible for half the production of heavy elements above iron on the periodic table. Masses are one of the most important nuclear physics ingredients that go into calculations of r-process nucleosynthesis as they enter into the calculations of reaction rates, decay rates, branching ratios and Q-values. We explore the impact of uncertainties in three nuclear mass models on r-process abundances by performing global monte carlo simulations. We show that root-mean-square (rms errors of current mass models are large so that current r-process predictions are insufficient in predicting features found in solar residuals and in r-process enhanced metal poor stars. We conclude that the reduction of global rms errors below 100 keV will allow for more robust r-process predictions.

  10. Implementation of the Business Process Modelling Notation (BPMN) in the modelling of anatomic pathology processes.

    Science.gov (United States)

    Rojo, Marcial García; Rolón, Elvira; Calahorra, Luis; García, Felix Oscar; Sánchez, Rosario Paloma; Ruiz, Francisco; Ballester, Nieves; Armenteros, María; Rodríguez, Teresa; Espartero, Rafael Martín

    2008-07-15

    Process orientation is one of the essential elements of quality management systems, including those in use in healthcare. Business processes in hospitals are very complex and variable. BPMN (Business Process Modelling Notation) is a user-oriented language specifically designed for the modelling of business (organizational) processes. Previous experiences of the use of this notation in the processes modelling within the Pathology in Spain or another country are not known. We present our experience in the elaboration of the conceptual models of Pathology processes, as part of a global programmed surgical patient process, using BPMN. With the objective of analyzing the use of BPMN notation in real cases, a multidisciplinary work group was created, including software engineers from the Dep. of Technologies and Information Systems from the University of Castilla-La Mancha and health professionals and administrative staff from the Hospital General de Ciudad Real. The work in collaboration was carried out in six phases: informative meetings, intensive training, process selection, definition of the work method, process describing by hospital experts, and process modelling. The modelling of the processes of Anatomic Pathology is presented using BPMN. The presented subprocesses are those corresponding to the surgical pathology examination of the samples coming from operating theatre, including the planning and realization of frozen studies. The modelling of Anatomic Pathology subprocesses has allowed the creation of an understandable graphical model, where management and improvements are more easily implemented by health professionals.

  11. Energy deposition model for low-energy electrons (10-10 000 eV) in air

    International Nuclear Information System (INIS)

    Roldan, A.; Perez, J.M.; Williart, A.; Blanco, F.; Garcia, G.

    2004-01-01

    An energy deposition model for electrons in air that can be useful in microdosimetric applications is presented in this study. The model is based on a Monte Carlo simulation of the single electron scattering processes that can take place with the molecular constituents of the air in the energy range 10-10 000 eV. The input parameters for this procedure have been the electron scattering cross sections, both differential and integral. These parameters were calculated using a model potential method which describes the electron scattering with the molecular constituent of air. The reliability of the calculated integral cross section values has been evaluated by comparison with direct total electron scattering cross-section measurements performed by us in a transmission beam experiment. Experimental energy loss spectra for electrons in air have been used as probability distribution functions to define the electron energy loss in single collision events. The resulting model has been applied to simulate the electron transport through a gas cell containing air at different pressures and the results have been compared with those observed in the experiments. Finally, as an example of its applicability to dosimetric issues, the energy deposition of 10 000 eV by means of successive collisions in a free air chamber has been simulated

  12. modeling grinding modeling grinding processes as micro processes

    African Journals Online (AJOL)

    eobe

    industrial precision grinding processes are cylindrical, center less and ... Several model shave been proposed and used to study grinding ..... grinding force for the two cases were 9.07237N/mm ..... International Journal of Machine Tools &.

  13. Non-equilibrium processes in p + Ag collisions at GeV energies

    Science.gov (United States)

    Fidelus, M.; Filges, D.; Goldenbaum, F.; Jarczyk, L.; Kamys, B.; Kistryn, M.; Kistryn, St.; Kozik, E.; Kulessa, P.; Machner, H.; Magiera, A.; Piskor-Ignatowicz, B.; Pysz, K.; Rudy, Z.; Sharma, Sushil K.; Siudak, R.; Wojciechowski, M.; PISA Collaboration

    2017-12-01

    The double differential spectra d2σ /d Ω d E of p , d , t , 3,4,6He, 6,7,8,9Li, 7,9,10Be, and 10,11,12B were measured at seven scattering angles, 15.6∘, 20∘, 35∘, 50∘, 65∘, 80∘, and 100∘, in the laboratory system for proton induced reactions on a silver target. Measurements were done for three proton energies: 1.2, 1.9, and 2.5 GeV. The experimental data were compared to calculations performed by means of two-step theoretical microscopic models. The first step of the reaction was described by the intranuclear cascade model incl4.6 and the second one by four different models (ABLA07, GEM2, gemini++, and SMM) using their standard parameter settings. Systematic deviations of the data from predictions of the models were observed. The deviations were especially large for the forward scattering angles and for the kinetic energy of emitted particles in the range from about 50 to 150 MeV. This suggests that some important non-equilibrium mechanism is lacking in the present day microscopic models of proton-nucleus collisions in the studied beam energy range.

  14. Search for the standard model Higgs boson in $e^+ e^-$ collisions at $\\sqrt{s}$=161, 170 and 172 GeV

    CERN Document Server

    Barate, R; Décamp, D; Ghez, P; Goy, C; Lees, J P; Lucotte, A; Minard, M N; Nief, J Y; Pietrzyk, B; Casado, M P; Chmeissani, M; Comas, P; Crespo, J M; Delfino, M C; Fernández, E; Fernández-Bosman, M; Garrido, L; Juste, A; Martínez, M; Merino, G; Miquel, R; Mir, L M; Padilla, C; Park, I C; Pascual, A; Perlas, J A; Riu, I; Sánchez, F; Teubert, F; Colaleo, A; Creanza, D; De Palma, M; Gelao, G; Iaselli, Giuseppe; Maggi, G; Maggi, M; Marinelli, N; Nuzzo, S; Ranieri, A; Raso, G; Ruggieri, F; Selvaggi, G; Silvestris, L; Tempesta, P; Tricomi, A; Zito, G; Huang, X; Lin, J; Ouyang, Q; Wang, T; Xie, Y; Xu, R; Xue, S; Zhang, J; Zhang, L; Zhao, W; Abbaneo, D; Alemany, R; Bazarko, A O; Becker, U; Bright-Thomas, P G; Cattaneo, M; Cerutti, F; Dissertori, G; Drevermann, H; Forty, Roger W; Frank, M; Hagelberg, R; Hansen, J B; Harvey, J; Janot, P; Jost, B; Kneringer, E; Knobloch, J; Lehraus, Ivan; Lutters, G; Mato, P; Minten, Adolf G; Moneta, L; Pacheco, A; Pusztaszeri, J F; Ranjard, F; Rizzo, G; Rolandi, Luigi; Rousseau, D; Schlatter, W D; Schmitt, M; Schneider, O; Tejessy, W; Tomalin, I R; Wachsmuth, H W; Wagner, A; Ajaltouni, Ziad J; Barrès, A; Boyer, C; Falvard, A; Ferdi, C; Gay, P; Guicheney, C; Henrard, P; Jousset, J; Michel, B; Monteil, S; Montret, J C; Pallin, D; Perret, P; Podlyski, F; Proriol, J; Rosnet, P; Rossignol, J M; Fearnley, Tom; Hansen, J D; Hansen, J R; Hansen, P H; Nilsson, B S; Rensch, B; Wäänänen, A; Daskalakis, G; Kyriakis, A; Markou, C; Simopoulou, Errietta; Vayaki, Anna; Blondel, A; Brient, J C; Machefert, F P; Rougé, A; Rumpf, M; Valassi, Andrea; Videau, H L; Focardi, E; Parrini, G; Zachariadou, K; Cavanaugh, R J; Corden, M; Georgiopoulos, C H; Hühn, T; Jaffe, D E; Antonelli, A; Bencivenni, G; Bologna, G; Bossi, F; Campana, P; Capon, G; Casper, David William; Chiarella, V; Felici, G; Laurelli, P; Mannocchi, G; Murtas, F; Murtas, G P; Passalacqua, L; Pepé-Altarelli, M; Curtis, L; Dorris, S J; Halley, A W; Knowles, I G; Lynch, J G; O'Shea, V; Raine, C; Scarr, J M; Smith, K; Teixeira-Dias, P; Thompson, A S; Thomson, E; Thomson, F; Turnbull, R M; Buchmüller, O L; Dhamotharan, S; Geweniger, C; Graefe, G; Hanke, P; Hansper, G; Hepp, V; Kluge, E E; Putzer, A; Sommer, J; Tittel, K; Werner, S; Wunsch, M; Beuselinck, R; Binnie, David M; Cameron, W; Dornan, Peter J; Girone, M; Goodsir, S M; Martin, E B; Morawitz, P; Moutoussi, A; Nash, J; Sedgbeer, J K; Spagnolo, P; Stacey, A M; Williams, M D; Ghete, V M; Girtler, P; Kuhn, D; Rudolph, G; Betteridge, A P; Bowdery, C K; Colrain, P; Crawford, G; Finch, A J; Foster, F; Hughes, G; Jones, R W L; Sloan, Terence; Whelan, E P; Williams, M I; Hoffmann, C; Jakobs, K; Kleinknecht, K; Quast, G; Renk, B; Rohne, E; Sander, H G; Van Gemmeren, P; Zeitnitz, C; Aubert, Jean-Jacques; Benchouk, C; Bonissent, A; Bujosa, G; Carr, J; Coyle, P; Diaconu, C A; Ealet, A; Fouchez, D; Konstantinidis, N P; Leroy, O; Motsch, F; Payre, P; Talby, M; Sadouki, A; Thulasidas, M; Tilquin, A; Trabelsi, K; Aleppo, M; Antonelli, M; Ragusa, F; Berlich, R; Blum, Walter; Büscher, V; Dietl, H; Ganis, G; Gotzhein, C; Kroha, H; Lütjens, G; Lutz, Gerhard; Männer, W; Moser, H G; Richter, R H; Rosado-Schlosser, A; Schael, S; Settles, Ronald; Seywerd, H C J; Saint-Denis, R; Stenzel, H; Wiedenmann, W; Wolf, G; Boucrot, J; Callot, O; Chen, S; Cordier, A; Davier, M; Duflot, L; Grivaz, J F; Heusse, P; Höcker, A; Jacholkowska, A; Jacquet, M; Kado, M; Kim, D W; Le Diberder, F R; Lefrançois, J; Lutz, A M; Nikolic, I A; Schune, M H; Serin, L; Simion, S; Tournefier, E; Veillet, J J; Videau, I; Zerwas, D; Azzurri, P; Bagliesi, G; Bettarini, S; Bozzi, C; Calderini, G; Ciulli, V; Dell'Orso, R; Fantechi, R; Ferrante, I; Giassi, A; Gregorio, A; Ligabue, F; Lusiani, A; Marrocchesi, P S; Messineo, A; Palla, Fabrizio; Sanguinetti, G; Sciabà, A; Steinberger, Jack; Tenchini, Roberto; Vannini, C; Venturi, A; Verdini, P G; Blair, G A; Bryant, L M; Chambers, J T; Gao, Y; Green, M G; Medcalf, T; Perrodo, P; Strong, J A; Von Wimmersperg-Töller, J H; Botterill, David R; Clifft, R W; Edgecock, T R; Haywood, S; Maley, P; Norton, P R; Thompson, J C; Wright, A E; Bloch-Devaux, B; Colas, P; Fabbro, B; Kozanecki, Witold; Lançon, E; Lemaire, M C; Locci, E; Pérez, P; Rander, J; Renardy, J F; Rosowsky, A; Roussarie, A; Schuller, J P; Schwindling, J; Trabelsi, A; Vallage, B; Black, S N; Dann, J H; Kim, H Y; Litke, A M; McNeil, M A; Taylor, G; Booth, C N; Boswell, R; Brew, C A J; Cartwright, S L; Combley, F; Kelly, M S; Lehto, M H; Newton, W M; Reeve, J; Thompson, L F; Affholderbach, K; Böhrer, A; Brandt, S; Cowan, G D; Foss, J; Grupen, Claus; Saraiva, P; Smolik, L; Stephan, F; Apollonio, M; Bosisio, L; Della Marina, R; Giannini, G; Gobbo, B; Musolino, G; Pütz, J; Rothberg, J E; Wasserbaech, S R; Williams, R W; Armstrong, S R; Charles, E; Elmer, P; Ferguson, D P S; González, S; Greening, T C; Hayes, O J; Hu, H; Jin, S; McNamara, P A; Nachtman, J M; Nielsen, J; Orejudos, W; Pan, Y B; Saadi, Y; Scott, I J; Walsh, J; Wu Sau Lan; Wu, X; Yamartino, J M; Zobernig, G

    1997-01-01

    The reaction e+e- -> HZ is used to search for the Standard Model Higgs boson. The data sample consists of integrated luminosities of 10.9pb-1 1.1pb-1 and 9.5pb-1 collected by the ALEPH experiment at LEP during 1996, at centre-of-mass energies of 161, 170 and 172GeV, respectively. No candidate events were found, in agreement with the expected background of 0.84 events from all Standard Model processes. This search results in a 95%C.L. lower limit on the Higgs boson mass of 69.4GeV. When combined with earlier ALEPH searches performed at energies at and around the Z peak, this limit increases to 70.7GeV

  15. Search for the Standard Model Higgs boson in e+e- collisions at sqrt(s)=161, 170 and 172GeV

    Science.gov (United States)

    ALEPH Collaboration; Barate, R.; Buskulic, D.; Decamp, D.; Ghez, P.; Goy, C.; Lees, J.-P.; Lucotte, A.; Minard, M.-N.; Nief, J.-Y.; Pietrzyk, B.; Casado, M. P.; Chmeissani, M.; Comas, P.; Crespo, J. M.; Delfino,, M.; Fernandez, E.; Fernandez-Bosman, M.; Garrido, Ll.; Juste, A.; Martinez, M.; Merino, G.; Miquel, R.; Mir, Ll. M.; Padilla, C.; Park, I. C.; Pascual, A.; Perlas, J. A.; Riu, I.; Sanchez, F.; Teubert, F.; Colaleo, A.; Creanza, D.; de Palma, M.; Gelao, G.; Iaselli, G.; Maggi, G.; Maggi, M.; Marinelli, N.; Nuzzo, S.; Ranieri, A.; Raso, G.; Ruggieri, F.; Selvaggi, G.; Silvestris, L.; Tempesta, P.; Tricomi, A.; Zito, G.; Huang, X.; Lin, J.; Ouyang, Q.; Wang, T.; Xie, Y.; Xu, R.; Xue, S.; Zhang, J.; Zhang, L.; Zhao, W.; Abbaneo, D.; Alemany, R.; Bazarko, A. O.; Becker, U.; Bright-Thomas, P.; Cattaneo, M.; Cerutti, F.; Dissertori, G.; Drevermann, H.; Forty, R. W.; Frank, M.; Hagelberg, R.; Hansen, J. B.; Harvey, J.; Janot, P.; Jost, B.; Kneringer, E.; Knobloch, J.; Lehraus, I.; Lutters, G.; Mato, P.; Minten, A.; Moneta, L.; Pacheco, A.; Pusztaszeri, J.-F.; Ranjard, F.; Rizzo, G.; Rolandi, L.; Rousseau, D.; Schlatter, D.; Schmitt, M.; Schneider, O.; Tejessy, W.; Tomalin, I. R.; Wachsmuth, H.; Wagner, A.; Ajaltouni, Z.; Barrès, A.; Boyer, C.; Falvard, A.; Ferdi, C.; Gay, P.; Guicheney, C.; Henrard, P.; Jousset, J.; Michel, B.; Monteil, S.; Montret, J.-C.; Pallin, D.; Perret, P.; Podlyski, F.; Proriol, J.; Rosnet, P.; Rossignol, J.-M.; Fearnley, T.; Hansen, J. D.; Hansen, J. R.; Hansen, P. H.; Nilsson, B. S.; Rensch, B.; Wäänänen, A.; Daskalakis, G.; Kyriakis, A.; Markou, C.; Simopoulou, E.; Vayaki, A.; Blondel, A.; Brient, J. C.; Machefert, F.; Rougé, A.; Rumpf, M.; Valassi, A.; Videau, H.; Focardi, E.; Parrini, G.; Zachariadou, K.; Cavanaugh, R.; Corden, M.; Georgiopoulos, C.; Huehn, T.; Jaffe, D. E.; Antonelli, A.; Bencivenni, G.; Bologna, G.; Bossi, F.; Campana, P.; Capon, G.; Casper, D.; Chiarella, V.; Felici, G.; Laurelli, P.; Mannocchi, G.; Murtas, F.; Murtas, G. P.; Passalacqua, L.; Pepe-Altarelli, M.; Curtis, L.; Dorris, S. J.; Halley, A. W.; Knowles, I. G.; Lynch, J. G.; O'Shea, V.; Raine, C.; Scarr, J. M.; Smith, K.; Teixeira-Dias, P.; Thompson, A. S.; Thomson, E.; Thomson, F.; Turnbull, R. M.; Buchmüller, O.; Dhamotharan, S.; Geweniger, C.; Graefe, G.; Hanke, P.; Hansper, G.; Hepp, V.; Kluge, E. E.; Putzer, A.; Sommer, J.; Tittel, K.; Werner, S.; Wunsch, M.; Beuselinck, R.; Binnie, D. M.; Cameron, W.; Dornan, P. J.; Girone, M.; Goodsir, S.; Martin, E. B.; Morawitz, P.; Moutoussi, A.; Nash, J.; Sedgbeer, J. K.; Spagnolo, P.; Stacey, A. M.; Williams, M. D.; Ghete, V. M.; Girtler, P.; Kuhn, D.; Rudolph, G.; Betteridge, A. P.; Bowdery, C. K.; Colrain, P.; Crawford, G.; Finch, A. J.; Foster, F.; Hughes, G.; Jones, R. W.; Sloan, T.; Whelan, E. P.; Williams, M. I.; Hoffmann, C.; Jakobs, K.; Kleinknecht, K.; Quast, G.; Renk, B.; Rohne, E.; Sander, H.-G.; van Gemmeren, P.; Zeitnitz, C.; Aubert, J. J.; Benchouk, C.; Bonissent, A.; Bujosa, G.; Carr, J.; Coyle, P.; Diaconu, C.; Ealet, A.; Fouchez, D.; Konstantinidis, N.; Leroy, O.; Motsch, F.; Payre, P.; Talby, M.; Sadouki, A.; Thulasidas, M.; Tilquin, A.; Trabelsi, K.; Aleppo, M.; Antonelli, M.; Ragusa, F.; Berlich, R.; Blum, W.; Büscher, V.; Dietl, H.; Ganis, G.; Gotzhein, C.; Kroha, H.; Lütjens, G.; Lutz, G.; Männer, W.; Moser, H.-G.; Richter, R.; Rosado-Schlosser, A.; Schael, S.; Settles, R.; Seywerd, H.; St. Denis, R.; Stenzel, H.; Wiedenmann, W.; Wolf, G.; Boucrot, J.; Callot, O.; Chen, S.; Cordier, A.; Davier, M.; Duflot, L.; Grivaz, J.-F.; Heusse, Ph.; Höcker, A.; Jacholkowska, A.; Jacquet, M.; Kado, M.; Kim, D. W.; Le Diberder, F.; Lefrançois, J.; Lutz, A.-M.; Nikolic, I.; Schune, M.-H.; Serin, L.; Simion, S.; Tournefier, E.; Veillet, J.-J.; Videau, I.; Zerwas, D.; Azzurri, P.; Bagliesi, G.; Bettarini, S.; Bozzi, C.; Calderini, G.; Ciulli, V.; dell'Orso, R.; Fantechi, R.; Ferrante, I.; Giassi, A.; Gregorio, A.; Ligabue, F.; Lusiani, A.; Marrocchesi, P. S.; Messineo, A.; Palla, F.; Sanguinetti, G.; Sciabà, A.; Steinberger, J.; Tenchini, R.; Vannini, C.; Venturi, A.; Verdini, P. G.; Blair, G. A.; Bryant, L. M.; Chambers, J. T.; Gao, Y.; Green, M. G.; Medcalf, T.; Perrodo, P.; Strong, J. A.; von Wimmersperg-Toeller, J. H.; Botterill, D. R.; Clifft, R. W.; Edgecock, T. R.; Haywood, S.; Maley, P.; Norton, P. R.; Thompson, J. C.; Wright, A. E.; Bloch-Devaux, B.; Colas, P.; Fabbro, B.; Kozanecki, W.; Lançon, E.; Lemaire, M. C.; Locci, E.; Perez, P.; Rander, J.; Renardy, J.-F.; Rosowsky, A.; Roussarie, A.; Schuller, J.-P.; Schwindling, J.; Trabelsi, A.; Vallage, B.; Black, S. N.; Dann, J. H.; Kim, H. Y.; Litke, A. M.; McNeil, M. A.; Taylor, G.; Booth, C. N.; Boswell, R.; Brew, C. A. J.; Cartwright, S.; Combley, F.; Kelly, M. S.; Lehto, M.; Newton, W. M.; Reeve, J.; Thompson, L. F.; Affholderbach, K.; Böhrer, A.; Brandt, S.; Cowan, G.; Foss, J.; Grupen, C.; Saraiva, P.; Smolik, L.; Stephan, F.; Apollonio, M.; Bosisio, L.; della Marina, R.; Giannini, G.; Gobbo, B.; Musolino, G.; Putz, J.; Rothberg, J.; Wasserbaech, S.; Williams, R. W.; Armstrong, S. R.; Charles, E.; Elmer, P.; Ferguson, D. P. S.; González, S.; Greening, T. C.; Hayes, O. J.; Hu, H.; Jin, S.; McNamara, P. A., III; Nachtman, J. M.; Nielsen, J.; Orejudos, W.; Pan, Y. B.; Saadi, Y.; Scott, I. J.; Walsh, J.; Wu, Sau Lan; Wu, X.; Yamartino, J. M.; Zobernig, G.

    1997-10-01

    The reaction e+e--->HZ is used to search for the Standard Model Higgs boson. The data sample consists of integrated luminosities of 10.9pb-1, 1.1pb-1, and 9.5pb-1 collected by the ALEPH experiment at LEP during 1996, at centre-of-mass energies of 161, 170 and 172GeV, respectively. No candidate events were found, in agreement with the expected background of 0.84 events from all Standard Model processes. This search results in a 95% C.L. lower limit on the Higgs boson mass of 69.4GeV/c2. When combined with earlier ALEPH searches performed at energies at and around the Z peak, this limit increases to 70.7GeV/c2.

  16. Die design and process optimization of die cast V6 engine blocks

    Directory of Open Access Journals (Sweden)

    Henry Hu

    2005-02-01

    Full Text Available The use of aluminum, particularly for engine blocks, has grown considerably in the past ten years, and continues to rise in the automotive industry. In order to enhance the quality and engineering functionality of die cast engine blocks, die design and processes have to be optimized. In this study, a computer simulation software, MAGMAsoft, as an advanced tool for optimizing die design and casting process, was emplooyed to virtually visualize cavity filling and patterns of a V6 engine block. The original die design and process was simulated first to establish a baseline. A reality check was used to verify the predicted results. Then, the die modification with a different unner system was made by using a CAD software, Unigraphics (UG. The simulation on combinations of the modified die design and revised process was performed to examine the effect of die modification and process change on flow filling of V6 engine blocks. The simulated prediction indicateds that the enhancement of cavity filling due to the die and process modification minimizeds the occurrence of defects during casting, and consequently improves the quality of blocks. The results of mechanical testing show a significant increase in fatigue strengths, and a moderately improvement on tensile properties for the blocks die cast with the new die design and prpocess in comparison with those produced by the original ones.

  17. Modeling of column apparatus processes

    CERN Document Server

    Boyadjiev, Christo; Boyadjiev, Boyan; Popova-Krumova, Petya

    2016-01-01

    This book presents a new approach for the modeling of chemical and interphase mass transfer processes in industrial column apparatuses, using convection-diffusion and average-concentration models. The convection-diffusion type models are used for a qualitative analysis of the processes and to assess the main, small and slight physical effects, and then reject the slight effects. As a result, the process mechanism can be identified. It also introduces average concentration models for quantitative analysis, which use the average values of the velocity and concentration over the cross-sectional area of the column. The new models are used to analyze different processes (simple and complex chemical reactions, absorption, adsorption and catalytic reactions), and make it possible to model the processes of gas purification with sulfur dioxide, which form the basis of several patents.

  18. Investigating the Process of Process Modeling with Eye Movement Analysis

    OpenAIRE

    Pinggera, Jakob; Furtner, Marco; Martini, Markus; Sachse, Pierre; Reiter, Katharina; Zugal, Stefan; Weber, Barbara

    2015-01-01

    Research on quality issues of business process models has recently begun to explore the process of creating process models by analyzing the modeler's interactions with the modeling environment. In this paper we aim to complement previous insights on the modeler's modeling behavior with data gathered by tracking the modeler's eye movements when engaged in the act of modeling. We present preliminary results and outline directions for future research to triangulate toward a more comprehensive un...

  19. What makes process models understandable?

    NARCIS (Netherlands)

    Mendling, J.; Reijers, H.A.; Cardoso, J.; Alonso, G.; Dadam, P.; Rosemann, M.

    2007-01-01

    Despite that formal and informal quality aspects are of significant importance to business process modeling, there is only little empirical work reported on process model quality and its impact factors. In this paper we investigate understandability as a proxy for quality of process models and focus

  20. Modeling the Object-Oriented Software Process: OPEN and the Unified Process

    NARCIS (Netherlands)

    van den Berg, Klaas; Aksit, Mehmet; van den Broek, P.M.

    A short introduction to software process modeling is presented, particularly object-oriented modeling. Two major industrial process models are discussed: the OPEN model and the Unified Process model. In more detail, the quality assurance in the Unified Process tool (formally called Objectory) is

  1. Search for the Standard Model Higgs Boson in $e^+ e^-$ Collisions at $\\sqrt{s}$ = 161-172 GeV

    CERN Document Server

    Ackerstaff, K; Allison, J; Altekamp, N; Anderson, K J; Anderson, S; Arcelli, S; Asai, S; Axen, D A; Azuelos, Georges; Ball, A H; Barberio, E; Barlow, R J; Bartoldus, R; Batley, J Richard; Baumann, S; Bechtluft, J; Beeston, C; Behnke, T; Bell, A N; Bell, K W; Bella, G; Bentvelsen, Stanislaus Cornelius Maria; Bethke, Siegfried; Biebel, O; Biguzzi, A; Bird, S D; Blobel, Volker; Bloodworth, Ian J; Bloomer, J E; Bobinski, M; Bock, P; Bonacorsi, D; Boutemeur, M; Bouwens, B T; Braibant, S; Brigliadori, L; Brown, R M; Burckhart, Helfried J; Burgard, C; Bürgin, R; Capiluppi, P; Carnegie, R K; Carter, A A; Carter, J R; Chang, C Y; Charlton, D G; Chrisman, D; Clarke, P E L; Cohen, I; Conboy, J E; Cooke, O C; Cuffiani, M; Dado, S; Dallapiccola, C; Dallavalle, G M; Davis, R; De Jong, S; del Pozo, L A; Desch, Klaus; Dienes, B; Dixit, M S; do Couto e Silva, E; Doucet, M; Duchovni, E; Duckeck, G; Duerdoth, I P; Eatough, D; Edwards, J E G; Estabrooks, P G; Evans, H G; Evans, M; Fabbri, Franco Luigi; Fanti, M; Faust, A A; Fiedler, F; Fierro, M; Fischer, H M; Fleck, I; Folman, R; Fong, D G; Foucher, M; Fürtjes, A; Futyan, D I; Gagnon, P; Gary, J W; Gascon, J; Gascon-Shotkin, S M; Geddes, N I; Geich-Gimbel, C; Geralis, T; Giacomelli, G; Giacomelli, P; Giacomelli, R; Gibson, V; Gibson, W R; Gingrich, D M; Glenzinski, D A; Goldberg, J; Goodrick, M J; Gorn, W; Grandi, C; Gross, E; Grunhaus, Jacob; Gruwé, M; Hajdu, C; Hanson, G G; Hansroul, M; Hapke, M; Hargrove, C K; Hart, P A; Hartmann, C; Hauschild, M; Hawkes, C M; Hawkings, R; Hemingway, Richard J; Herndon, M; Herten, G; Heuer, R D; Hildreth, M D; Hill, J C; Hillier, S J; Hobson, P R; Homer, R James; Honma, A K; Horváth, D; Hossain, K R; Howard, R; Hüntemeyer, P; Hutchcroft, D E; Igo-Kemenes, P; Imrie, D C; Ingram, M R; Ishii, K; Jawahery, A; Jeffreys, P W; Jeremie, H; Jimack, Martin Paul; Joly, A; Jones, C R; Jones, G; Jones, M; Jost, U; Jovanovic, P; Junk, T R; Karlen, D A; Kartvelishvili, V G; Kawagoe, K; Kawamoto, T; Kayal, P I; Keeler, Richard K; Kellogg, R G; Kennedy, B W; Kirk, J; Klier, A; Kluth, S; Kobayashi, T; Kobel, M; Koetke, D S; Kokott, T P; Kolrep, M; Komamiya, S; Kress, T; Krieger, P; Von Krogh, J; Kyberd, P; Lafferty, G D; Lahmann, R; Lai, W P; Lanske, D; Lauber, J; Lautenschlager, S R; Layter, J G; Lazic, D; Lee, A M; Lefebvre, E; Lellouch, Daniel; Letts, J; Levinson, L; Lloyd, S L; Loebinger, F K; Long, G D; Losty, Michael J; Ludwig, J; Macchiolo, A; MacPherson, A L; Mannelli, M; Marcellini, S; Markus, C; Martin, A J; Martin, J P; Martínez, G; Mashimo, T; Mättig, P; McDonald, W J; McKenna, J A; McKigney, E A; McMahon, T J; McPherson, R A; Meijers, F; Menke, S; Merritt, F S; Mes, H; Meyer, J; Michelini, Aldo; Mikenberg, G; Miller, D J; Mincer, A; Mir, R; Mohr, W; Montanari, A; Mori, T; Morii, M; Müller, U; Mihara, S; Nagai, K; Nakamura, I; Neal, H A; Nellen, B; Nisius, R; O'Neale, S W; Oakham, F G; Odorici, F; Ögren, H O; Oh, A; Oldershaw, N J; Oreglia, M J; Orito, S; Pálinkás, J; Pásztor, G; Pater, J R; Patrick, G N; Patt, J; Pearce, M J; Pérez-Ochoa, R; Petzold, S; Pfeifenschneider, P; Pilcher, J E; Pinfold, J L; Plane, D E; Poffenberger, P R; Poli, B; Posthaus, A; Rees, D L; Rigby, D; Robertson, S; Robins, S A; Rodning, N L; Roney, J M; Rooke, A M; Ros, E; Rossi, A M; Routenburg, P; Rozen, Y; Runge, K; Runólfsson, O; Ruppel, U; Rust, D R; Rylko, R; Sachs, K; Saeki, T; Sarkisyan-Grinbaum, E; Sbarra, C; Schaile, A D; Schaile, O; Scharf, F; Scharff-Hansen, P; Schenk, P; Schieck, J; Schleper, P; Schmitt, B; Schmitt, S; Schöning, A; Schröder, M; Schultz-Coulon, H C; Schumacher, M; Schwick, C; Scott, W G; Shears, T G; Shen, B C; Shepherd-Themistocleous, C H; Sherwood, P; Siroli, G P; Sittler, A; Skillman, A; Skuja, A; Smith, A M; Snow, G A; Sobie, Randall J; Söldner-Rembold, S; Springer, R W; Sproston, M; Stephens, K; Steuerer, J; Stockhausen, B; Stoll, K; Strom, D; Szymanski, P; Tafirout, R; Talbot, S D; Tanaka, S; Taras, P; Tarem, S; Teuscher, R; Thiergen, M; Thomson, M A; Von Törne, E; Towers, S; Trigger, I; Trócsányi, Z L; Tsur, E; Turcot, A S; Turner-Watson, M F; Utzat, P; Van Kooten, R; Verzocchi, M; Vikas, P; Vokurka, E H; Voss, H; Wäckerle, F; Wagner, A; Ward, C P; Ward, D R; Watkins, P M; Watson, A T; Watson, N K; Wells, P S; Wermes, N; White, J S; Wilkens, B; Wilson, G W; Wilson, J A; Wolf, G; Wyatt, T R; Yamashita, S; Yekutieli, G; Zacek, V; Zer-Zion, D

    1998-01-01

    This paper describes a search for the Standard Model Higgs boson using data from e^+e^- collisions collected at center-of-mass energies of 161, 170 and 172 GeV by the OPAL detector at LEP. The data collected at these energies correspond to integrated luminosities of 10.0, 1.0 and 9.4 pb^-1, respectively. The search is sensitive to the main final states from the process in which the Higgs boson is produced in association with a fermion anti-fermion pair, namely four jets, two jets with missing energy, and two jets produced together with a pair of electron, muon or tau leptons. One candidate event is observed, in agreement with the Standard Model background expectation. In combination with previous OPAL searches at center-of-mass energies close to the Z^0 resonance and the revised previous OPAL searches at 161 GeV, we derive a lower limit of 69.4 GeV for the mass of the Standard Model Higgs boson at the 95% confidence level.

  2. Modeling the Object-Oriented Software Process: OPEN and the Unified Process

    OpenAIRE

    van den Berg, Klaas; Aksit, Mehmet; van den Broek, P.M.

    1999-01-01

    A short introduction to software process modeling is presented, particularly object-oriented modeling. Two major industrial process models are discussed: the OPEN model and the Unified Process model. In more detail, the quality assurance in the Unified Process tool (formally called Objectory) is reviewed.

  3. High Energy pp Elastic Scattering in Condensate Enclosed Chiral Bag Model and TOTEM Elastic Measurements at LHC at 7 TeV

    CERN Document Server

    Islam, M M

    2013-01-01

    We study high energy $\\small{\\rm{pp}}$ and $\\small{\\rm{\\bar {p}p}}$ elastic scattering in the TeV region based on an effective field theory model of the proton. We phenomenologically investigate the main processes underlying elastic scattering and quantitatively describe the measured elastic d$\\small{\\sigma}$/dt at energies 7.0 TeV (LHC $\\small{\\rm{pp}}$), 1.96 TeV (Tevatron $\\small{\\rm{\\bar {p}p}}$), and 0.630 TeV (SPS $\\small{\\rm{\\bar {p}p}}$). Finally, we give our prediction for $\\small{\\rm{pp}}$ elastic d$\\small{\\sigma}$/dt at 14 TeV that will be measured by the TOTEM Collaboration.

  4. Model unspecific search in CMS. First results at 13 TeV

    Energy Technology Data Exchange (ETDEWEB)

    Roemer, Jonas; Albert, Andreas; Duchardt, Deborah; Hebbeker, Thomas; Knutzen, Simon; Lieb, Jonas; Meyer, Arnd; Pook, Tobias [III. Physikalisches Institut A, RWTH Aachen University (Germany)

    2016-07-01

    Following an upgrade in center of mass energy from √(s) = 8 TeV to 13 TeV, the LHC delivered first proton-proton collisions at this unprecedented energy in 2015. The CMS experiment recorded data corresponding to an integrated luminosity of 3.7 fb{sup -1}. Since many theoretical models predict signal cross sections to increase strongly with the center of mass energy, the data taken at √(s) = 13 TeV are competitive to the previous data taking period even with a lower recorded integrated luminosity. The Model Unspecific Search in CMS (MUSiC) searches for physics beyond the standard model independent of theoretical models. Using an automatic method, kinematic distributions of the data are compared with the standard model expectation in every final state. Therefore, MUSiC reduces the chance of overlooking new physics, since even distributions not covered by dedicated analyses are investigated. This talk outlines changes to the analysis made necessary by the increased center of mass energy and first results with lepton triggered events.

  5. Developing Worksheet (LKS) Base on Process Skills in Curriculum 2013 at Elementary School Grade IV,V,VI

    Science.gov (United States)

    Subhan, M.; Oktolita, N.; Kn, M.

    2018-04-01

    The Lacks of students' skills in the learning process is due to lacks of exercises in the form of LKS. In the curriculum of 2013, there is no LKS as a companion to improve the students' skills. In order to solve those problem, it is necessary to develop LKS based on process skills as a teaching material to improve students' process skills. The purpose of this study is to develop LKS Process Skills based elementary school grade IV, V, VI which is integrated by process skill. The development of LKS can be used to develop the thematic process skills of elementary school students grade IV, V, VI based on curriculum 2013. The expected long-term goal is to produce teaching materials LKS Process Skill based of Thematic learning that is able to develop the process skill of elementary school students grade IV, V, VI. This development research refers to the steps developed by Borg & Gall (1983). The development process is carried out through 10 stages: preliminary research and gathering information, planning, draft development, initial test (limited trial), first product revision, final trial (field trial), product operational revision, Desemination and implementation. The limited subject of the this research is the students of SDN in Dharmasraya grade IV, V, VI. The field trial subjects in the experimental class are the students of SDN Dharmasraya grade IV, V, VI who have implemented the curriculum 2013. The data are collected by using LKS validation sheets, process skill observation sheets, and Thematic learning test (pre-test And post-test). The result of LKS development on the validity score is 81.70 (very valid), on practical score is 83.94 (very practical), and on effectiveness score is 86.67 (very effective). In the trial step the use of LKS using One Group Pretest-Posttest Design research design. The purpose of this trial is to know the effectiveness level of LKS result of development for improving the process skill of students in grade IV, V, and VI of elementary

  6. Multi-enzyme Process Modeling

    DEFF Research Database (Denmark)

    Andrade Santacoloma, Paloma de Gracia

    are affected (in a positive or negative way) by the presence of the other enzymes and compounds in the media. In this thesis the concept of multi-enzyme in-pot term is adopted for processes that are carried out by the combination of enzymes in a single reactor and implemented at pilot or industrial scale...... features of the process and provides the information required to structure the process model by using a step-by-step procedure with the required tools and methods. In this way, this framework increases efficiency of the model development process with respect to time and resources needed (fast and effective....... In this way the model parameters that drives the main dynamic behavior can be identified and thus a better understanding of this type of processes. In order to develop, test and verify the methodology, three case studies were selected, specifically the bi-enzyme process for the production of lactobionic acid...

  7. Process model repositories and PNML

    NARCIS (Netherlands)

    Hee, van K.M.; Post, R.D.J.; Somers, L.J.A.M.; Werf, van der J.M.E.M.; Kindler, E.

    2004-01-01

    Bringing system and process models together in repositories facilitates the interchange of model information between modelling tools, and allows the combination and interlinking of complementary models. Petriweb is a web application for managing such repositories. It supports hierarchical process

  8. Model-based software process improvement

    Science.gov (United States)

    Zettervall, Brenda T.

    1994-01-01

    The activities of a field test site for the Software Engineering Institute's software process definition project are discussed. Products tested included the improvement model itself, descriptive modeling techniques, the CMM level 2 framework document, and the use of process definition guidelines and templates. The software process improvement model represents a five stage cyclic approach for organizational process improvement. The cycles consist of the initiating, diagnosing, establishing, acting, and leveraging phases.

  9. a Fast Segmentation Algorithm for C-V Model Based on Exponential Image Sequence Generation

    Science.gov (United States)

    Hu, J.; Lu, L.; Xu, J.; Zhang, J.

    2017-09-01

    For the island coastline segmentation, a fast segmentation algorithm for C-V model method based on exponential image sequence generation is proposed in this paper. The exponential multi-scale C-V model with level set inheritance and boundary inheritance is developed. The main research contributions are as follows: 1) the problems of the "holes" and "gaps" are solved when extraction coastline through the small scale shrinkage, low-pass filtering and area sorting of region. 2) the initial value of SDF (Signal Distance Function) and the level set are given by Otsu segmentation based on the difference of reflection SAR on land and sea, which are finely close to the coastline. 3) the computational complexity of continuous transition are successfully reduced between the different scales by the SDF and of level set inheritance. Experiment results show that the method accelerates the acquisition of initial level set formation, shortens the time of the extraction of coastline, at the same time, removes the non-coastline body part and improves the identification precision of the main body coastline, which automates the process of coastline segmentation.

  10. Technical Area V (TA-V) transformation project close-out report.

    Energy Technology Data Exchange (ETDEWEB)

    2010-07-01

    Sandia National Laboratories (SNL) Technical Area V (TA-V) has provided unique nuclear experimental environments for decades. The technologies tested in TA-V facilities have furthered the United States Nuclear Weapons program and has contributed to the national energy and homeland security mission. The importance of TA-V working efficiently to produce an attractive and effective platform for experiments should not be underestimated. Throughout its brief history, TA-V has evolved to address multiple and diverse sets of requirements. These requirements evolved over many years; however, the requirements had not been managed nor communicated comprehensively or effectively. A series of programmatic findings over several years of external audits was evidence of this downfall. Today, these same requirements flow down through a new TA-V management system that produces consistently applied and reproducible approaches to work practices. In 2008, the TA-V department managers assessed the state of TA-V services and work activities to understand how to improve customer interfaces, stakeholders perceptions, and workforce efficiencies. The TA-V management team initiated the TA-V Transformation Project after they deemed the pre-June 2008 operational model to be ineffective in managing work and in providing integrated, continuous improvement to TA-V processes. This report summarizes the TA-V Transformation Project goals, activities, and accomplishments.

  11. Kinetic modelling of a diesel-polluted clayey soil bioremediation process

    Energy Technology Data Exchange (ETDEWEB)

    Fernández, Engracia Lacasa; Merlo, Elena Moliterni [Chemical Engineering Department, Research Institute for Chemical and Environmental Technology (ITQUIMA), University of Castilla La Mancha, 13071 Ciudad Real (Spain); Mayor, Lourdes Rodríguez [National Institute for Hydrogen Research, C/Fernando el Santo, 13500 Puertollano (Spain); Camacho, José Villaseñor, E-mail: jose.villasenor@uclm.es [Chemical Engineering Department, Research Institute for Chemical and Environmental Technology (ITQUIMA), University of Castilla La Mancha, 13071 Ciudad Real (Spain)

    2016-07-01

    A mathematical model is proposed to describe a diesel-polluted clayey soil bioremediation process. The reaction system under study was considered a completely mixed closed batch reactor, which initially contacted a soil matrix polluted with diesel hydrocarbons, an aqueous liquid-specific culture medium and a microbial inoculation. The model coupled the mass transfer phenomena and the distribution of hydrocarbons among four phases (solid, S; water, A; non-aqueous liquid, NAPL; and air, V) with Monod kinetics. In the first step, the model simulating abiotic conditions was used to estimate only the mass transfer coefficients. In the second step, the model including both mass transfer and biodegradation phenomena was used to estimate the biological kinetic and stoichiometric parameters. In both situations, the model predictions were validated with experimental data that corresponded to previous research by the same authors. A correct fit between the model predictions and the experimental data was observed because the modelling curves captured the major trends for the diesel distribution in each phase. The model parameters were compared to different previously reported values found in the literature. Pearson correlation coefficients were used to show the reproducibility level of the model. - Highlights: • A mathematical model is proposed to describe a soil bioremediation process. • The model couples mass transfer phenomena among phases with biodegradation. • Model predictions were validated with previous data reported by the authors. • A correct fit and correlation coefficients were observed.

  12. Kinetic modelling of a diesel-polluted clayey soil bioremediation process

    International Nuclear Information System (INIS)

    Fernández, Engracia Lacasa; Merlo, Elena Moliterni; Mayor, Lourdes Rodríguez; Camacho, José Villaseñor

    2016-01-01

    A mathematical model is proposed to describe a diesel-polluted clayey soil bioremediation process. The reaction system under study was considered a completely mixed closed batch reactor, which initially contacted a soil matrix polluted with diesel hydrocarbons, an aqueous liquid-specific culture medium and a microbial inoculation. The model coupled the mass transfer phenomena and the distribution of hydrocarbons among four phases (solid, S; water, A; non-aqueous liquid, NAPL; and air, V) with Monod kinetics. In the first step, the model simulating abiotic conditions was used to estimate only the mass transfer coefficients. In the second step, the model including both mass transfer and biodegradation phenomena was used to estimate the biological kinetic and stoichiometric parameters. In both situations, the model predictions were validated with experimental data that corresponded to previous research by the same authors. A correct fit between the model predictions and the experimental data was observed because the modelling curves captured the major trends for the diesel distribution in each phase. The model parameters were compared to different previously reported values found in the literature. Pearson correlation coefficients were used to show the reproducibility level of the model. - Highlights: • A mathematical model is proposed to describe a soil bioremediation process. • The model couples mass transfer phenomena among phases with biodegradation. • Model predictions were validated with previous data reported by the authors. • A correct fit and correlation coefficients were observed.

  13. Separation of hadronic W and Z decays in the CLIC_ILD and the CLICdet detector models at 1.4 and 3TeV

    CERN Document Server

    AUTHOR|(SzGeCERN)793139; Roloff, Philipp Gerhard; Strom, Lars Rickard; Weber, Matthias Artur

    2017-01-01

    A study of the W and Z separation was performed for the CLIC_ILD and the CLICdet detector models for the proposed Compact Linear Collider (CLIC). Comparisons were done for fully-hadronic WW and ZZ events at the collision energies of 1.4 and 3 TeV. Particle flow objects are reconstructed using a full simulation of the events including relevant beam-induced background processes. Several different collections of particles, with varying level of background suppression, were compared for each of the detector models and optimal jet clustering parameters were found in each case, resulting in the best separation of the W and Z mass peaks. The CLICdet detector model performs similar to CLIC_ILD with an achieved jet mass separation of around 1.6 $\\sigma$ at 1.4 TeV and 1.3 $\\sigma$ at 3 TeV. For both detector models we achieve a better separation at 1.4 TeV when comparing dijet masses rather than large-R jet masses. At 3 TeV jets with a radius around R=0.5 perform similarly well as dijets.

  14. Using Unified Modelling Language (UML) as a process-modelling technique for clinical-research process improvement.

    Science.gov (United States)

    Kumarapeli, P; De Lusignan, S; Ellis, T; Jones, B

    2007-03-01

    The Primary Care Data Quality programme (PCDQ) is a quality-improvement programme which processes routinely collected general practice computer data. Patient data collected from a wide range of different brands of clinical computer systems are aggregated, processed, and fed back to practices in an educational context to improve the quality of care. Process modelling is a well-established approach used to gain understanding and systematic appraisal, and identify areas of improvement of a business process. Unified modelling language (UML) is a general purpose modelling technique used for this purpose. We used UML to appraise the PCDQ process to see if the efficiency and predictability of the process could be improved. Activity analysis and thinking-aloud sessions were used to collect data to generate UML diagrams. The UML model highlighted the sequential nature of the current process as a barrier for efficiency gains. It also identified the uneven distribution of process controls, lack of symmetric communication channels, critical dependencies among processing stages, and failure to implement all the lessons learned in the piloting phase. It also suggested that improved structured reporting at each stage - especially from the pilot phase, parallel processing of data and correctly positioned process controls - should improve the efficiency and predictability of research projects. Process modelling provided a rational basis for the critical appraisal of a clinical data processing system; its potential maybe underutilized within health care.

  15. WW + jet at 14 and 100 TeV

    Energy Technology Data Exchange (ETDEWEB)

    Campbell, John [Fermilab; Miller, David [Glasgow U.; Robens, Tania [Dresden, Tech. U.

    2016-11-05

    In the current LHC run, an accurate understanding of Standard Model processes is extremely important. Processes including electroweak gauge bosons serve as standard candles for SM measurements, and equally constitute important backgrounds for Beyond-the-Standard Model (BSM) searches. We present here the next-to-leading order (NLO) QCD virtual contributions to W+W- + jet in an analytic format obtained through unitarity methods. We present results for the full process using the Monte Carlo event generator MCFM, and discuss total as well as differential cross-sections for the LHC with 14 TeV center-of-mass energy, as well as a future 100 TeV proton-proton machine.

  16. Absolute cross sections for the multielectron processes in 15 keV I10++rare gas collisions

    International Nuclear Information System (INIS)

    Nakamura, N.; Currell, F.J.; Danjo, A.; Kimura, M.; Matsumoto, A.; Ohtani, S.; Sakaue, H.A.; Sakurai, M.; Tawara, H.; Watanabe, H.; Yamada, I.; Yoshino, M.

    1995-01-01

    We have experimentally determined the absolute cross sections for total charge transfer (σ q ), j electron transfer (σ j q ), i electron capture (σ q,q-i ) and each reaction process (σ j q,q-i ) in 15 keV I 10+ -Ne, Ar, Kr and Xe collisions. The branching ratios were determined by the coincidence measurements between charge changing projectile and recoil ions. The electron capture cross sections were measured by the initial growth rate method. The experimental results for total and j electron transfer cross sections were compared with the predictions of the extended classical over-barrier model (ECBM). (orig.)

  17. Developing engineering processes through integrated modelling of product and process

    DEFF Research Database (Denmark)

    Nielsen, Jeppe Bjerrum; Hvam, Lars

    2012-01-01

    This article aims at developing an operational tool for integrated modelling of product assortments and engineering processes in companies making customer specific products. Integrating a product model in the design of engineering processes will provide a deeper understanding of the engineering...... activities as well as insight into how product features affect the engineering processes. The article suggests possible ways of integrating models of products with models of engineering processes. The models have been tested and further developed in an action research study carried out in collaboration...... with a major international engineering company....

  18. VIP : A Visual Editor and Compiler for v-Promela

    OpenAIRE

    Kamel, Moataz; Leue, Stefan

    2001-01-01

    We describe the Visual Interface to Promela (VIP) tool that we have recently implemented. VIP supports the visual editing and maintenance of v-Promela models. v-Promela is a visual, object-oriented extension to Promela, the input language to the Spin model checker. We introduce the v-Promela notation as supported by the VIP editor, discuss Promela code generation, and describe the process of property validation for the resulting models. Our discussion centers around two case studies, a call p...

  19. UML in business process modeling

    Directory of Open Access Journals (Sweden)

    Bartosz Marcinkowski

    2013-03-01

    Full Text Available Selection and proper application of business process modeling methods and techniques have a significant impact on organizational improvement capabilities as well as proper understanding of functionality of information systems that shall support activity of the organization. A number of business process modeling notations were popularized in practice in recent decades. Most significant of the notations include Business Process Modeling Notation (OMG BPMN and several Unified Modeling Language (OMG UML extensions. In this paper, the assessment whether one of the most flexible and strictly standardized contemporary business process modeling notations, i.e. Rational UML Profile for Business Modeling, enable business analysts to prepare business models that are all-embracing and understandable by all the stakeholders. After the introduction, methodology of research is discussed. Section 2 presents selected case study results. The paper is concluded with a summary.

  20. Main modelling features of the ASTEC V2.1 major version

    International Nuclear Information System (INIS)

    Chatelard, P.; Belon, S.; Bosland, L.; Carénini, L.; Coindreau, O.; Cousin, F.; Marchetto, C.; Nowack, H.; Piar, L.; Chailan, L.

    2016-01-01

    Highlights: • Recent modelling improvements of the ASTEC European severe accident code are outlined. • Key new physical models now available in the ASTEC V2.1 major version are described. • ASTEC progress towards a multi-design reactor code is illustrated for BWR and PHWR. • ASTEC strong link with the on-going EC CESAM FP7 project is emphasized. • Main remaining modelling issues (on which IRSN efforts are now directing) are given. - Abstract: A new major version of the European severe accident integral code ASTEC, developed by IRSN with some GRS support, was delivered in November 2015 to the ASTEC worldwide community. Main modelling features of this V2.1 version are summarised in this paper. In particular, the in-vessel coupling technique between the reactor coolant system thermal-hydraulics module and the core degradation module has been strongly re-engineered to remove some well-known weaknesses of the former V2.0 series. The V2.1 version also includes new core degradation models specifically addressing BWR and PHWR reactor types, as well as several other physical modelling improvements, notably on reflooding of severely damaged cores, Zircaloy oxidation under air atmosphere, corium coolability during corium concrete interaction and source term evaluation. Moreover, this V2.1 version constitutes the back-bone of the CESAM FP7 project, which final objective is to further improve ASTEC for use in Severe Accident Management analysis of the Gen.II–III nuclear power plants presently under operation or foreseen in near future in Europe. As part of this European project, IRSN efforts to continuously improve both code numerical robustness and computing performances at plant scale as well as users’ tools are being intensified. Besides, ASTEC will continue capitalising the whole knowledge on severe accidents phenomenology by progressively keeping physical models at the state of the art through a regular feed-back from the interpretation of the current and

  1. Automatic extraction of process categories from process model collections

    NARCIS (Netherlands)

    Malinova, M.; Dijkman, R.M.; Mendling, J.; Lohmann, N.; Song, M.; Wohed, P.

    2014-01-01

    Many organizations build up their business process management activities in an incremental way. As a result, there is no overarching structure defined at the beginning. However, as business process modeling initiatives often yield hundreds to thousands of process models, there is a growing need for

  2. Biosphere Process Model Report

    Energy Technology Data Exchange (ETDEWEB)

    J. Schmitt

    2000-05-25

    To evaluate the postclosure performance of a potential monitored geologic repository at Yucca Mountain, a Total System Performance Assessment (TSPA) will be conducted. Nine Process Model Reports (PMRs), including this document, are being developed to summarize the technical basis for each of the process models supporting the TSPA model. These reports cover the following areas: (1) Integrated Site Model; (2) Unsaturated Zone Flow and Transport; (3) Near Field Environment; (4) Engineered Barrier System Degradation, Flow, and Transport; (5) Waste Package Degradation; (6) Waste Form Degradation; (7) Saturated Zone Flow and Transport; (8) Biosphere; and (9) Disruptive Events. Analysis/Model Reports (AMRs) contain the more detailed technical information used to support TSPA and the PMRs. The AMRs consists of data, analyses, models, software, and supporting documentation that will be used to defend the applicability of each process model for evaluating the postclosure performance of the potential Yucca Mountain repository system. This documentation will ensure the traceability of information from its source through its ultimate use in the TSPA-Site Recommendation (SR) and in the National Environmental Policy Act (NEPA) analysis processes. The objective of the Biosphere PMR is to summarize (1) the development of the biosphere model, and (2) the Biosphere Dose Conversion Factors (BDCFs) developed for use in TSPA. The Biosphere PMR does not present or summarize estimates of potential radiation doses to human receptors. Dose calculations are performed as part of TSPA and will be presented in the TSPA documentation. The biosphere model is a component of the process to evaluate postclosure repository performance and regulatory compliance for a potential monitored geologic repository at Yucca Mountain, Nevada. The biosphere model describes those exposure pathways in the biosphere by which radionuclides released from a potential repository could reach a human receptor

  3. Biosphere Process Model Report

    International Nuclear Information System (INIS)

    Schmitt, J.

    2000-01-01

    To evaluate the postclosure performance of a potential monitored geologic repository at Yucca Mountain, a Total System Performance Assessment (TSPA) will be conducted. Nine Process Model Reports (PMRs), including this document, are being developed to summarize the technical basis for each of the process models supporting the TSPA model. These reports cover the following areas: (1) Integrated Site Model; (2) Unsaturated Zone Flow and Transport; (3) Near Field Environment; (4) Engineered Barrier System Degradation, Flow, and Transport; (5) Waste Package Degradation; (6) Waste Form Degradation; (7) Saturated Zone Flow and Transport; (8) Biosphere; and (9) Disruptive Events. Analysis/Model Reports (AMRs) contain the more detailed technical information used to support TSPA and the PMRs. The AMRs consists of data, analyses, models, software, and supporting documentation that will be used to defend the applicability of each process model for evaluating the postclosure performance of the potential Yucca Mountain repository system. This documentation will ensure the traceability of information from its source through its ultimate use in the TSPA-Site Recommendation (SR) and in the National Environmental Policy Act (NEPA) analysis processes. The objective of the Biosphere PMR is to summarize (1) the development of the biosphere model, and (2) the Biosphere Dose Conversion Factors (BDCFs) developed for use in TSPA. The Biosphere PMR does not present or summarize estimates of potential radiation doses to human receptors. Dose calculations are performed as part of TSPA and will be presented in the TSPA documentation. The biosphere model is a component of the process to evaluate postclosure repository performance and regulatory compliance for a potential monitored geologic repository at Yucca Mountain, Nevada. The biosphere model describes those exposure pathways in the biosphere by which radionuclides released from a potential repository could reach a human receptor

  4. Conceptual models of information processing

    Science.gov (United States)

    Stewart, L. J.

    1983-01-01

    The conceptual information processing issues are examined. Human information processing is defined as an active cognitive process that is analogous to a system. It is the flow and transformation of information within a human. The human is viewed as an active information seeker who is constantly receiving, processing, and acting upon the surrounding environmental stimuli. Human information processing models are conceptual representations of cognitive behaviors. Models of information processing are useful in representing the different theoretical positions and in attempting to define the limits and capabilities of human memory. It is concluded that an understanding of conceptual human information processing models and their applications to systems design leads to a better human factors approach.

  5. Microstructure evolution and grain refinement of Ti-6Al-4V alloy by laser shock processing

    Energy Technology Data Exchange (ETDEWEB)

    Ren, X.D., E-mail: renxd@mail.ujs.edu.cn [Department of Mechanical Engineering, Jiangsu University, Zhenjiang, 212013 (China); Research Center of Fluid Machinery Engineering and Technical, Jiangsu University, Zhenjiang, 212013 (China); Zhou, W.F.; Liu, F.F.; Ren, Y.P. [Department of Mechanical Engineering, Jiangsu University, Zhenjiang, 212013 (China); Yuan, S.Q. [Research Center of Fluid Machinery Engineering and Technical, Jiangsu University, Zhenjiang, 212013 (China); Ren, N.F.; Xu, S.D.; Yang, T. [Department of Mechanical Engineering, Jiangsu University, Zhenjiang, 212013 (China)

    2016-02-15

    Graphical abstract: The grain refinement process of Ti-6Al-4V alloy under LSP: (a) LDD in original grains; (b) Dislocations in β phase; (c) DTIs in α phase; (d) DTs transform into DCs; (e) DWs develop into sub-GBs; (f) GR accomplishes. - Highlights: • LSP could repair the surface defects and reduce the surface roughness. • Microstructure evolution of α phase in Ti-6Al-4V alloy processed by LSP is distinct from β phase. • Multidirectional twin intersections and subgrain boundaries are the main mechanism of grain refinement of Ti-6Al-4V alloy. • Grain refinement process of the Ti-6Al-4V alloy was illustrated. - Abstract: Microstructure evolution and grain refinement of Ti-6Al-4V alloy after laser shock processing (LSP) are systematically investigated in this paper. Laser shock waves were induced by a Q-switched Nd:YAG laser system operated with a wave-length of 1064 nm and 10 ns pulse width. The microstructures of LSP samples were characterized by scanning electron microscopy (SEM) and transmission electron microscope (TEM). Present results indicate that the surface hardness of samples subjected to LSP impacts has significantly improved. Multidirectional twin intersections and dislocation movements lead to grain subdivision in α phase with ultra-high plastic deformation. High-density dislocations are found in β phase. Multidirectional twin intersections and division of sub-grain boundaries play an important role in the grain refinement of Ti-6Al-4V alloy under LSP loading conditions.

  6. P-barp and pp elastic scattering from 10 GeV to 1000 GeV centre-of-mass energy

    International Nuclear Information System (INIS)

    Islam, M.M.; Fearnley, T.; Guillaud, J.P.

    1984-01-01

    Antiproton-proton and proton-proton elastic scattering are studied simultaneously over the energy range √s approx. (10-1000) GeV in a nucleon valence core model proposed earlier. The scattering is described as primarily due to two processes: diffraction and hard scattering. The latter originates from the scattering of a nucleon core off another core. Destructive interference between the two processes produces dips in p-barp and pp differential cross-sections. As energy increases beyond the ISR range (√s = (23-62) GeV), the dips get filled up, and eventually transform into shoulders or breaks at collider energies. Differences between p-barp and pp differential cross-sections persist even at collider energies. Comparison with ISR data shows that the model provides a quantitative description of pp elastic scattering in this energy range. Predictions of p-barp and pp differential cross-sections at future collider energies √s = 800 and 2000 GeV are given. In order to distinguish between competing models, need for measuring the p-barp differential cross-section at the ISR and SPS collider in the abs (t)-range (0.5-2.0) (GeV) 2 is stressed

  7. anti pp and pp elastic scattering from 10 GeV to 1000 GeV centre-of-mass energy

    Energy Technology Data Exchange (ETDEWEB)

    Islam, M.M. (Connecticut Univ., Storrs (USA). Dept. of Physics); Fearnley, T. (University Coll., London (UK). Dept. of Physics and Astronomy); Guillaud, J.P. (Grenoble-1 Univ., 74 - Annecy (France). Lab. de Physique des Particules)

    1984-06-21

    Antiproton-proton and proton-proton elastic scattering are studied simultaneously over the energy range ..sqrt..anti s approx.= (10/1000) GeV in a nucleon valence core model proposed earlier. The scattering is described as primarily due to two processes: diffraction and hard scattering. The latter originates from the scattering of a nucleon core off another core. Destructive interference between the two processes produces dips in anti pp and pp differential cross-sections. As energy increases beyond the ISR range (..sqrt..anti s = (23/62) GeV), the dips get filled up, and eventually transform into shoulders or breaks at collider energies. Differences between anti pp and pp differential cross-sections persist even at collider energies. Comparison with ISR data shows that the model provides a quantitative description of pp elastic scattering in this energy range. Predictions of anti pp and pp differential cross-sections at future collider energies ..sqrt..s = 800 and 2000 GeV are given. In order to distinguish between competing models, need for measuring the anti pp differential cross-section at the ISR and SPS collider in the vertical stroketvertical stroke-range (0.5/2.0) (GeV)/sup 2/ is stressed.

  8. P-barp and pp elastic scattering from 10 GeV to 1000 GeV centre-of-mass energy

    Energy Technology Data Exchange (ETDEWEB)

    Islam, M.M. (Connecticut Univ., Storrs (USA). Dept. of Physics); Fearnley, T. (University Coll., London (UK). Dept. of Physics and Astronomy); Guillaud, J.P. (L.A.P.P. - BP909, 74019 Annecy-Le-Vieux Cedex, France)

    1984-06-21

    Antiproton-proton and proton-proton elastic scattering are studied simultaneously over the energy range ..sqrt..s approx. (10-1000) GeV in a nucleon valence core model proposed earlier. The scattering is described as primarily due to two processes: diffraction and hard scattering. The latter originates from the scattering of a nucleon core off another core. Destructive interference between the two processes produces dips in p-barp and pp differential cross-sections. As energy increases beyond the ISR range (..sqrt..s = (23-62) GeV), the dips get filled up, and eventually transform into shoulders or breaks at collider energies. Differences between p-barp and pp differential cross-sections persist even at collider energies. Comparison with ISR data shows that the model provides a quantitative description of pp elastic scattering in this energy range. Predictions of p-barp and pp differential cross-sections at future collider energies ..sqrt..s = 800 and 2000 GeV are given. In order to distinguish between competing models, need for measuring the p-barp differential cross-section at the ISR and SPS collider in the abs (t)-range (0.5-2.0) (GeV)/sup 2/ is stressed.

  9. Life Cycle V and V Process for Hardware Description Language Programs of Programmable Logic Device-based Instrumentation and Control Systems

    International Nuclear Information System (INIS)

    Cha, K. H.; Lee, D. Y.

    2010-01-01

    Programmable Logic Device (PLD), especially Complex PLD (CPLD) or Field Programmable Logic Array (FPGA), has been growing in interest in nuclear Instrumentation and Control (I and C) applications. PLD has been applied to replace an obsolete analog device or old-fashioned microprocessor, or to develop digital controller, subsystem or overall system on hardware aspects. This is the main reason why the PLD-based I and C design provides higher flexibility than the analog-based one, and the PLD-based I and C systems shows better real-time performance than the processor-based I and C systems. Due to the development of the PLD-based I and C systems, their nuclear qualification has been issued in the nuclear industry. Verification and Validation (V and V) is one of necessary qualification activities when a Hardware Description Language (HDL) is used to implement functions of the PLD-based I and C systems. The life cycle V and V process, described in this paper, has been defined as satisfying the nuclear V and V requirements, and it has been applied to verify Correctness, Completeness, and Consistency (3C) among design outputs in a safety-grade programmable logic controller and a safety-critical data communication system. Especially, software engineering techniques such as the Fagan Inspection, formal verification, simulated verification and automated testing have been defined for the life cycle V and V tasks of behavioral, structural, and physical design in VHDL

  10. Numerical modelling of the V-J combinations of the T cell receptor TRA/TRD locus.

    Directory of Open Access Journals (Sweden)

    Florence Thuderoz

    2010-02-01

    Full Text Available T-Cell antigen Receptor (TR repertoire is generated through rearrangements of V and J genes encoding alpha and beta chains. The quantification and frequency for every V-J combination during ontogeny and development of the immune system remain to be precisely established. We have addressed this issue by building a model able to account for Valpha-Jalpha gene rearrangements during thymus development of mice. So we developed a numerical model on the whole TRA/TRD locus, based on experimental data, to estimate how Valpha and Jalpha genes become accessible to rearrangements. The progressive opening of the locus to V-J gene recombinations is modeled through windows of accessibility of different sizes and with different speeds of progression. Furthermore, the possibility of successive secondary V-J rearrangements was included in the modelling. The model points out some unbalanced V-J associations resulting from a preferential access to gene rearrangements and from a non-uniform partition of the accessibility of the J genes, depending on their location in the locus. The model shows that 3 to 4 successive rearrangements are sufficient to explain the use of all the V and J genes of the locus. Finally, the model provides information on both the kinetics of rearrangements and frequencies of each V-J associations. The model accounts for the essential features of the observed rearrangements on the TRA/TRD locus and may provide a reference for the repertoire of the V-J combinatorial diversity.

  11. Numerical Modelling Of The V-J Combinations Of The T Cell Receptor TRA/TRD Locus

    Science.gov (United States)

    Dariz, Aurélie; Baum, Thierry Pascal; Hierle, Vivien; Demongeot, Jacques; Marche, Patrice Noël; Jouvin-Marche, Evelyne

    2010-01-01

    T-Cell antigen Receptor (TR) repertoire is generated through rearrangements of V and J genes encoding α and β chains. The quantification and frequency for every V-J combination during ontogeny and development of the immune system remain to be precisely established. We have addressed this issue by building a model able to account for Vα-Jα gene rearrangements during thymus development of mice. So we developed a numerical model on the whole TRA/TRD locus, based on experimental data, to estimate how Vα and Jα genes become accessible to rearrangements. The progressive opening of the locus to V-J gene recombinations is modeled through windows of accessibility of different sizes and with different speeds of progression. Furthermore, the possibility of successive secondary V-J rearrangements was included in the modelling. The model points out some unbalanced V-J associations resulting from a preferential access to gene rearrangements and from a non-uniform partition of the accessibility of the J genes, depending on their location in the locus. The model shows that 3 to 4 successive rearrangements are sufficient to explain the use of all the V and J genes of the locus. Finally, the model provides information on both the kinetics of rearrangements and frequencies of each V-J associations. The model accounts for the essential features of the observed rearrangements on the TRA/TRD locus and may provide a reference for the repertoire of the V-J combinatorial diversity. PMID:20174554

  12. Improved model management with aggregated business process models

    NARCIS (Netherlands)

    Reijers, H.A.; Mans, R.S.; Toorn, van der R.A.

    2009-01-01

    Contemporary organizations invest much efforts in creating models of their business processes. This raises the issue of how to deal with large sets of process models that become available over time. This paper proposes an extension of Event-driven Process Chains, called the aggregate EPC (aEPC),

  13. Diversity, social inquiries, and epistemic virtues

    Directory of Open Access Journals (Sweden)

    Adler, Jonathan Eric

    2005-01-01

    Full Text Available A teoria das virtudes epistêmicas (VE sustenta que as virtudes dos agentes, tais como a imparcialidade ou a permeabilidade intelectual, ao invés de crenças específicas, devem estar no centro da avaliação epistêmica, e que os individuos que possuem essas virtudes estão mais bem-posicionados epistemicamente do que se não as tivessem, ou, pior ainda, do que se tivessem os vícios correspondentes: o preconceito, o dogmatismo, ou a impermeabilidade intelectual. Eu argumento que a teoria VE padece de um grave defeito, porque fracassa ao se ajustar à natureza social dos questionamentos (epistêmicos típicos. Esse e outros defeitos relacionados a esse infectam o paralelo que os teóricos VE traçam entre virtudes epistêmicas e morais. Ao prometer o incremento na proporção de crenças verdadeiras sobre crenças falsas, ou ignorância, as virtudes epistêmicas não podem desempenhar um papel paralelo àquele que Aristóteles reserva às virtudes morais ao prometer o incremento em nossa felicidade e no bem-estar da comunidade. A minha rota para essas críticas é feita das razões sobre por que os agentes (sociais devem buscar a obtenção de seus objetivos morais e epistêmicos diferentemente nos papéis que atribuem às virtudes

  14. Business process modeling in healthcare.

    Science.gov (United States)

    Ruiz, Francisco; Garcia, Felix; Calahorra, Luis; Llorente, César; Gonçalves, Luis; Daniel, Christel; Blobel, Bernd

    2012-01-01

    The importance of the process point of view is not restricted to a specific enterprise sector. In the field of health, as a result of the nature of the service offered, health institutions' processes are also the basis for decision making which is focused on achieving their objective of providing quality medical assistance. In this chapter the application of business process modelling - using the Business Process Modelling Notation (BPMN) standard is described. Main challenges of business process modelling in healthcare are the definition of healthcare processes, the multi-disciplinary nature of healthcare, the flexibility and variability of the activities involved in health care processes, the need of interoperability between multiple information systems, and the continuous updating of scientific knowledge in healthcare.

  15. Integrating textual and model-based process descriptions for comprehensive process search

    NARCIS (Netherlands)

    Leopold, Henrik; van der Aa, Han; Pittke, Fabian; Raffel, Manuel; Mendling, Jan; Reijers, Hajo A.

    2016-01-01

    Documenting business processes using process models is common practice in many organizations. However, not all process information is best captured in process models. Hence, many organizations complement these models with textual descriptions that specify additional details. The problem with this

  16. Exclusive data-based modeling of neutron-nuclear reactions below 20 MeV

    Science.gov (United States)

    Savin, Dmitry; Kosov, Mikhail

    2017-09-01

    We are developing CHIPS-TPT physics library for exclusive simulation of neutron-nuclear reactions below 20 MeV. Exclusive modeling reproduces each separate scattering and thus requires conservation of energy, momentum and quantum numbers in each reaction. Inclusive modeling reproduces only selected values while averaging over the others and imposes no such constraints. Therefore the exclusive modeling allows to simulate additional quantities like secondary particle correlations and gamma-lines broadening and avoid artificial fluctuations. CHIPS-TPT is based on the formerly included in Geant4 CHIPS library, which follows the exclusive approach, and extends it to incident neutrons with the energy below 20 MeV. The NeutronHP model for neutrons below 20 MeV included in Geant4 follows the inclusive approach like the well known MCNP code. Unfortunately, the available data in this energy region is mostly presented in ENDF-6 format and semi-inclusive. Imposing additional constraints on secondary particles complicates modeling but also allows to detect inconsistencies in the input data and to avoid errors that may remain unnoticed in inclusive modeling.

  17. Pair production of pions with symmetric momenta in the range 0.5 <= Psub(T) <= 2.0 GeV/c in 70-GeV p-p collisions

    International Nuclear Information System (INIS)

    Abramov, V.V.; Baldin, V.Yu.; Buzulutskov, A.F.

    1981-01-01

    The process of pion pair production is studied in the 70 GeV pp collisions. The invariant cross section slope of the pp → π + π - + X process as a function of transverse mompsub(T)entum is found to have a break near 1 GeV/c. Fitting the cross section by a sum of two exponents gives the values of powers (12.3+-0.9) (GeV/c) -1 and (8.7+-0.6) (GeV/c) -1 . The experimental points at psub(T)>=1 GeV/c are significantly higher than predictions based on hard scattering models such as quantum chromodynamics and constituent interchange model. The largest disagreement is discovered for calculations of the cross section in the framework of quantum chromodynamics [ru

  18. The Application of V&V within Reuse-Based Software Engineering

    Science.gov (United States)

    Addy, Edward

    1996-01-01

    Verification and Validation (V&V) is performed during application development for many systems, especially safety-critical and mission-critical systems. The V&V process is intended to discover errors as early as possible during the development process. Early discovery is important in order to minimize the cost and other impacts of correcting these errors. In reuse-based software engineering, decisions on the requirements, design and even implementation of domain assets can can be made prior to beginning development of a specific system. in order to bring the effectiveness of V&V to bear within reuse-based software engineering. V&V must be incorporated within the domain engineering process.

  19. Analysis of p-bar p scattering at 31 GeV and 62 GeV by the Chou-Yang model

    International Nuclear Information System (INIS)

    Padua, A.B. de; Covolan, R.J.M.; Souza Paes, J.T. de

    1988-01-01

    The p-bar p scattering is analysed at 31 GeV and 62 GeV energies for momentum transfers in the range O 2 . The experimental (dσ/dt)p-bar p values were fitted using a pure imaginary written as a sum of exponentials, that is, a(s,t)=a(s,O) σ n i=l α i e βit . Using the parameters obtained we have calculated the absorption constant K p-bar p the form factor and the mean square radius of the p-bar matter distribuition by the Chou-Yang model. These calculations reveal a ''dip'' around -t approx.= 1.3 (GeV/c) 2 at 31 GeV and 62 GeV. (author) [pt

  20. INC Model interpretation of the proton induced residual nuclide production cross sections below 2 GeV

    International Nuclear Information System (INIS)

    Divadeenam, M.; Ward, T.E.; Spergel, M.S.; Lakatos, S.; Manche, E.P.

    1991-01-01

    For the purposes of interpreting the abundances of various isotopes in meteorites or on lunar and planetary surfaces exposed to fragmentation by cosmic rays, Webber et al. recently reported the measured total elemental and isotopic cross sections with heavy ions as projectiles on H, He, and C targets with beam energies of 0.33 - 1.7 GeV/nucleon. We employ the INC model to predict the fragmentation of the heavy ions in a hydrogen target with the inverse reaction process: proton bombardment of a heavy-ion nucleus leading to spallation products. Charge-changing and mass-changing cross sections are calculated for proton bombardment of an 56 Fe target with beam energies ranging from 0.33 to 1.88 GeV. Total Z-changing and A-changing cross sections in the energy range 0.6 to 1.88 GeV are in excellent agreement with the corresponding experimental data of Webber et al. and Westfall at al., while the agreement below 0.6 GeV proton energy is not as good. The general trend of the Z-changing cross sections are reproduced by the model calculations at each proton incident energy. The interaction of 200-MeV protons with synthetic Stony Meteorite samples was undertaken to explain radionuclide production in a cosmic-ray environment. The BNL Linac 200-MeV-proton beam was used to irradiate synthetic Stony Meteorites to simulate cosmic-ray exposures corresponding to 6.4 and 16.4 million years. Each irradiated sample was analyzed with the help of a high-resolution gamma-ray spectrometer for long-lived radioisotopes. The intranuclear cascade code HETC was employed to simulate the 200-MeV proton bombardment on the meteorite samples to predict the radionuclides 7 Be, 22 Na, 46 Mn, and 56 Co produced in the experimental investigation

  1. Process modeling for Humanities: tracing and analyzing scientific processes

    OpenAIRE

    Hug , Charlotte; Salinesi , Camille; Deneckere , Rebecca; Lamasse , Stéphane

    2011-01-01

    International audience; This paper concerns epistemology and the understanding of research processes in Humanities, such as Archaeology. We believe that to properly understand research processes, it is essential to trace them. The collected traces depend on the process model established, which has to be as accurate as possible to exhaustively record the traces. In this paper, we briefly explain why the existing process models for Humanities are not sufficient to represent traces. We then pres...

  2. Modeling and process optimization of electrospinning of chitosan-collagen nanofiber by response surface methodology

    Science.gov (United States)

    Amiri, Nafise; Moradi, Ali; Abolghasem Sajjadi Tabasi, Sayyed; Movaffagh, Jebrail

    2018-04-01

    Chitosan-collagen composite nanofiber is of a great interest to researchers in biomedical fields. Since the electrospinning is the most popular method for nanofiber production, having a comprehensive knowledge of the electrospinning process is beneficial. Modeling techniques are precious tools for managing variables in the electrospinning process, prior to the more time- consuming and expensive experimental techniques. In this study, a central composite design of response surface methodology (RSM) was employed to develop a statistical model as well as to define the optimum condition for fabrication of chitosan-collagen nanofiber with minimum diameter. The individual and the interaction effects of applied voltage (10–25 kV), flow rate (0.5–1.5 mL h‑1), and needle to collector distance (15–25 cm) on the fiber diameter were investigated. ATR- FTIR and cell study were done to evaluate the optimized nanofibers. According to the RSM, a two-factor interaction (2FI) model was the most suitable model. The high regression coefficient value (R 2 ≥ 0.9666) of the fitted regression model and insignificant lack of fit (P = 0.0715) indicated that the model was highly adequate in predicting chitosan-collagen nanofiber diameter. The optimization process showed that the chitosan-collagen nanofiber diameter of 156.05 nm could be obtained in 9 kV, 0.2 ml h‑1, and 25 cm which was confirmed by experiment (155.92 ± 18.95 nm). The ATR-FTIR and cell study confirmed the structure and biocompatibility of the optimized membrane. The represented model could assist researchers in fabricating chitosan-collagen electrospun scaffolds with a predictable fiber diameter, and optimized chitosan-collagen nanofibrous mat could be a potential candidate for wound healing and tissue engineering.

  3. Photon+V measurements in ATLAS

    CERN Document Server

    Krasnopevtsev, Dimitrii; The ATLAS collaboration

    2017-01-01

    ATLAS measurements of multi-boson production processes involving isolated photons in proton–proton collisions at 8 TeV are summarized. Standard Model cross sections are measured with high precision and are compared to theoretical predictions. No deviations from Standard Model predictions are observed and limits are placed on parameters used to describe anomalous triple and quartic gauge-boson couplings.

  4. The EUCLID/V1 Integrated Code for Safety Assessment of Liquid Metal Cooled Fast Reactors. Part 1: Basic Models

    Science.gov (United States)

    Mosunova, N. A.

    2018-05-01

    The article describes the basic models included in the EUCLID/V1 integrated code intended for safety analysis of liquid metal (sodium, lead, and lead-bismuth) cooled fast reactors using fuel rods with a gas gap and pellet dioxide, mixed oxide or nitride uranium-plutonium fuel under normal operation, under anticipated operational occurrences and accident conditions by carrying out interconnected thermal-hydraulic, neutronics, and thermal-mechanical calculations. Information about the Russian and foreign analogs of the EUCLID/V1 integrated code is given. Modeled objects, equation systems in differential form solved in each module of the EUCLID/V1 integrated code (the thermal-hydraulic, neutronics, fuel rod analysis module, and the burnup and decay heat calculation modules), the main calculated quantities, and also the limitations on application of the code are presented. The article also gives data on the scope of functions performed by the integrated code's thermal-hydraulic module, using which it is possible to describe both one- and twophase processes occurring in the coolant. It is shown that, owing to the availability of the fuel rod analysis module in the integrated code, it becomes possible to estimate the performance of fuel rods in different regimes of the reactor operation. It is also shown that the models implemented in the code for calculating neutron-physical processes make it possible to take into account the neutron field distribution over the fuel assembly cross section as well as other features important for the safety assessment of fast reactors.

  5. Modeling nuclear processes by Simulink

    Energy Technology Data Exchange (ETDEWEB)

    Rashid, Nahrul Khair Alang Md, E-mail: nahrul@iium.edu.my [Faculty of Engineering, International Islamic University Malaysia, Jalan Gombak, Selangor (Malaysia)

    2015-04-29

    Modelling and simulation are essential parts in the study of dynamic systems behaviours. In nuclear engineering, modelling and simulation are important to assess the expected results of an experiment before the actual experiment is conducted or in the design of nuclear facilities. In education, modelling can give insight into the dynamic of systems and processes. Most nuclear processes can be described by ordinary or partial differential equations. Efforts expended to solve the equations using analytical or numerical solutions consume time and distract attention from the objectives of modelling itself. This paper presents the use of Simulink, a MATLAB toolbox software that is widely used in control engineering, as a modelling platform for the study of nuclear processes including nuclear reactor behaviours. Starting from the describing equations, Simulink models for heat transfer, radionuclide decay process, delayed neutrons effect, reactor point kinetic equations with delayed neutron groups, and the effect of temperature feedback are used as examples.

  6. Modeling nuclear processes by Simulink

    International Nuclear Information System (INIS)

    Rashid, Nahrul Khair Alang Md

    2015-01-01

    Modelling and simulation are essential parts in the study of dynamic systems behaviours. In nuclear engineering, modelling and simulation are important to assess the expected results of an experiment before the actual experiment is conducted or in the design of nuclear facilities. In education, modelling can give insight into the dynamic of systems and processes. Most nuclear processes can be described by ordinary or partial differential equations. Efforts expended to solve the equations using analytical or numerical solutions consume time and distract attention from the objectives of modelling itself. This paper presents the use of Simulink, a MATLAB toolbox software that is widely used in control engineering, as a modelling platform for the study of nuclear processes including nuclear reactor behaviours. Starting from the describing equations, Simulink models for heat transfer, radionuclide decay process, delayed neutrons effect, reactor point kinetic equations with delayed neutron groups, and the effect of temperature feedback are used as examples

  7. On the use of the plasma in III-V semiconductor processing

    Energy Technology Data Exchange (ETDEWEB)

    Bruno, G.; Capezzuto, P.; Losurdo, M. [C.N.R.-Centro di Studio per la Chimica dei Plasmi Dipartimento di Chimica-Universita di Bari via Orabona, 4-70126 Bari (Italy)

    1996-03-01

    The manufacture of usable devices based on III-V semiconductor materials is a complex process requiring epilayer growth, anisotropic etching, defect passivation, surface oxidation and substrate preparation processes. The combination of plasma based methods with metalorganic chemical vapor deposition (MOCVD) offers some real advantages: {ital in} {ital situ} production and preactivation of PH{sub 3} and sample preparation using H-atom. The detailed understanding and use of the plasma (using mass spectrometry, optical emission spectroscopy, laser reflectance interferometry and spectroscopic ellipsometry) as applied to InP material is discussed. {copyright} {ital 1996 American Institute of Physics.}

  8. Effect of V-Nd co-doping on phase transformation and grain growth process of TiO2

    Science.gov (United States)

    Khatun, Nasima; Amin, Ruhul; Anita, Sen, Somaditya

    2018-05-01

    The pure and V-Nd co-doped TiO2 samples are prepared by the modified sol-gel process. The phase formation is confirmed by XRD spectrum. Phase transformation is delayed in V-Nd co-doped TiO2 (TVN) samples compared to pure TiO2. The particle size is comparatively small in TVN samples at both the temperature 450 °C and 900 °C. Hence the effect of Nd doping is dominated over V doping in both phase transformation and grain growth process of TiO2.

  9. Sato Processes in Default Modeling

    DEFF Research Database (Denmark)

    Kokholm, Thomas; Nicolato, Elisa

    -change of a homogeneous Levy process. While the processes in these two classes share the same average behavior over time, the associated intensities exhibit very different properties. Concrete specifications are calibrated to data on the single names included in the iTraxx Europe index. The performances are compared......In reduced form default models, the instantaneous default intensity is classically the modeling object. Survival probabilities are then given by the Laplace transform of the cumulative hazard defined as the integrated intensity process. Instead, recent literature has shown a tendency towards...... specifying the cumulative hazard process directly. Within this framework we present a new model class where cumulative hazards are described by self-similar additive processes, also known as Sato processes. Furthermore we also analyze specifications obtained via a simple deterministic time...

  10. A FAST SEGMENTATION ALGORITHM FOR C-V MODEL BASED ON EXPONENTIAL IMAGE SEQUENCE GENERATION

    Directory of Open Access Journals (Sweden)

    J. Hu

    2017-09-01

    Full Text Available For the island coastline segmentation, a fast segmentation algorithm for C-V model method based on exponential image sequence generation is proposed in this paper. The exponential multi-scale C-V model with level set inheritance and boundary inheritance is developed. The main research contributions are as follows: 1 the problems of the "holes" and "gaps" are solved when extraction coastline through the small scale shrinkage, low-pass filtering and area sorting of region. 2 the initial value of SDF (Signal Distance Function and the level set are given by Otsu segmentation based on the difference of reflection SAR on land and sea, which are finely close to the coastline. 3 the computational complexity of continuous transition are successfully reduced between the different scales by the SDF and of level set inheritance. Experiment results show that the method accelerates the acquisition of initial level set formation, shortens the time of the extraction of coastline, at the same time, removes the non-coastline body part and improves the identification precision of the main body coastline, which automates the process of coastline segmentation.

  11. Computerized urea kinetic modeling to prescribe and monitor delivered Kt/V (pKt/V, dKt/V) in peritoneal dialysis. Fresenius Randomized Dialysis Prescriptions and Clinical Outcome Study (RDP/CO).

    Science.gov (United States)

    Gotch, F A; Lipps, B J; Keen, M L; Panlilio, F

    1996-01-01

    A computerized urea kinetic model of peritoneal urea transport (PACK-PD) has been developed and used to calculate prescription parameters which would result in the prescribed weekly peritoneal urea clearance (pKpt/V) required to achieve levels of weekly summed renal + peritoneal urea clearance (pKprt/ V) targeted at 1.75 and 2.16. Baseline kinetic data were obtained and analyzed with PACK-PD on 88 patients, and the program then used these data to calculate the required pKpt/V and subsequently the delivered Kpt/V (dKpt/V) from the dialysate collections. A total of 108 prescriptions were written and compared to dKpt/V measured over one to 24 months in the 88 patients. Both continuous ambulatory peritoneal dialysis and automated peritoneal dialysis (APD) were studied (APD consisted of PD+ with one or two diurnal and two to four nocturnal cycler exchanges). The correlation of dKpt/V to pKpt/V showed r = 0.93 with 95% confidence limits (CL) on agreement of +/-20% over a range of pKpt/V 0.52-2.55. The 95% CL on (dKpt/V-pKpt/V) were +/-0.30. We concluded: (1) that the prescription can be modeled as reliably in peritoneal dialysis as in hemodialysis (HD) where dKt/V and pKt/V agree to +/-25%, (2) that any individual weekly dKpt/V may vary as much as 0.3-0.4 from pKpt/V, and (3) that frequent measurement of dKpt/V and adjustment of pKpt/V as needed are required (as in HD) to control mean dKpt/V to within +/-10% of mean pKpt/V.

  12. Scalable approximate policies for Markov decision process models of hospital elective admissions.

    Science.gov (United States)

    Zhu, George; Lizotte, Dan; Hoey, Jesse

    2014-05-01

    To demonstrate the feasibility of using stochastic simulation methods for the solution of a large-scale Markov decision process model of on-line patient admissions scheduling. The problem of admissions scheduling is modeled as a Markov decision process in which the states represent numbers of patients using each of a number of resources. We investigate current state-of-the-art real time planning methods to compute solutions to this Markov decision process. Due to the complexity of the model, traditional model-based planners are limited in scalability since they require an explicit enumeration of the model dynamics. To overcome this challenge, we apply sample-based planners along with efficient simulation techniques that given an initial start state, generate an action on-demand while avoiding portions of the model that are irrelevant to the start state. We also propose a novel variant of a popular sample-based planner that is particularly well suited to the elective admissions problem. Results show that the stochastic simulation methods allow for the problem size to be scaled by a factor of almost 10 in the action space, and exponentially in the state space. We have demonstrated our approach on a problem with 81 actions, four specialities and four treatment patterns, and shown that we can generate solutions that are near-optimal in about 100s. Sample-based planners are a viable alternative to state-based planners for large Markov decision process models of elective admissions scheduling. Copyright © 2014 Elsevier B.V. All rights reserved.

  13. V-SIPAL - A VIRTUAL LABORATORY FOR SATELLITE IMAGE PROCESSING AND ANALYSIS

    Directory of Open Access Journals (Sweden)

    K. M. Buddhiraju

    2012-09-01

    Full Text Available In this paper a virtual laboratory for the Satellite Image Processing and Analysis (v-SIPAL being developed at the Indian Institute of Technology Bombay is described. v-SIPAL comprises a set of experiments that are normally carried out by students learning digital processing and analysis of satellite images using commercial software. Currently, the experiments that are available on the server include Image Viewer, Image Contrast Enhancement, Image Smoothing, Edge Enhancement, Principal Component Transform, Texture Analysis by Co-occurrence Matrix method, Image Indices, Color Coordinate Transforms, Fourier Analysis, Mathematical Morphology, Unsupervised Image Classification, Supervised Image Classification and Accuracy Assessment. The virtual laboratory includes a theory module for each option of every experiment, a description of the procedure to perform each experiment, the menu to choose and perform the experiment, a module on interpretation of results when performed with a given image and pre-specified options, bibliography, links to useful internet resources and user-feedback. The user can upload his/her own images for performing the experiments and can also reuse outputs of one experiment in another experiment where applicable. Some of the other experiments currently under development include georeferencing of images, data fusion, feature evaluation by divergence andJ-M distance, image compression, wavelet image analysis and change detection. Additions to the theory module include self-assessment quizzes, audio-video clips on selected concepts, and a discussion of elements of visual image interpretation. V-SIPAL is at the satge of internal evaluation within IIT Bombay and will soon be open to selected educational institutions in India for evaluation.

  14. Prediction of springback in V-die air bending process by using finite element method

    Directory of Open Access Journals (Sweden)

    Trzepiecinski Tomasz

    2017-01-01

    Full Text Available Springback phenomenon affects the dimensional and geometrical accuracy of the bent parts. The prediction of springback is a key problem in sheet metal forming. The aim of this paper is the numerical analysis of the possibility to predict the springback of anisotropic steel sheets. The experiments are conducted on 40 x 100 mm steel sheets. The mechanical properties of the sheet metals have been determined through uniaxial tensile tests of samples cut along three directions with respect to the rolling direction. The numerical model of air V-bending is built in finite element method (FEM based ABAQUS/Standard 2016.HF2 (Dassault Systemes Simulia Corp., USA program. The FEM results were verified by experimental investigations. The simulation model has taken into consideration material anisotropy and strain hardening phenomenon. The results of FEM simulations confirmed the ability of numerical prediction of springback amount. It was also found that the directional microstructure of the sheet metal resulted from rolling process affects the elastic-plastic deformation of the sheets through the sample width.

  15. Transferable tight-binding model for strained group IV and III-V materials and heterostructures

    Science.gov (United States)

    Tan, Yaohua; Povolotskyi, Michael; Kubis, Tillmann; Boykin, Timothy B.; Klimeck, Gerhard

    2016-07-01

    It is critical to capture the effect due to strain and material interface for device level transistor modeling. We introduce a transferable s p3d5s* tight-binding model with nearest-neighbor interactions for arbitrarily strained group IV and III-V materials. The tight-binding model is parametrized with respect to hybrid functional (HSE06) calculations for varieties of strained systems. The tight-binding calculations of ultrasmall superlattices formed by group IV and group III-V materials show good agreement with the corresponding HSE06 calculations. The application of the tight-binding model to superlattices demonstrates that the transferable tight-binding model with nearest-neighbor interactions can be obtained for group IV and III-V materials.

  16. Study of Various Processes with 160 A GeV Pb Beam

    CERN Multimedia

    2002-01-01

    % WA101 \\\\ \\\\ Ten modules of BP-1 glass detectors interleaved with various targets ranging from C to Pb were exposed to the 160~A~GeV~Pb beam in the November-December run of 1994 at CERN SPS. The experiment was carried out at normal incidence at a beam density of $\\sim$~600~cm$ ^- ^{2} $. The dimension of each plate of BP-1 glass was 50~mm~x~50~mm~x~1~mm. We etched the glass in 70\\% CH$ _{3} $SO$ _{3} $H at 50 $^{0}$C or in 48\\%~HF at room temperature. The charge threshold is found to be Z$ _{t} _{h} $ $\\sim$ 68 and 70 respectively. Using the automated scanning and measurement system developed at Berkeley, we have demonstrated that the charge resolution for Pb ions is $\\sigma _{Z} $~=~0.14 charge unit from a single measurement within a distance of 30~$\\mu$m. This excellent charge resolution allows us to make the proposed measurements of cross-sections for various processes. \\\\ \\\\We use this detector system to measure cross-sections for various processes in heavy ion collisions of 160~A~GeV~Pb with different t...

  17. Multiphysics modelling of manufacturing processes: A review

    DEFF Research Database (Denmark)

    Jabbari, Masoud; Baran, Ismet; Mohanty, Sankhya

    2018-01-01

    Numerical modelling is increasingly supporting the analysis and optimization of manufacturing processes in the production industry. Even if being mostly applied to multistep processes, single process steps may be so complex by nature that the needed models to describe them must include multiphysics...... the diversity in the field of modelling of manufacturing processes as regards process, materials, generic disciplines as well as length scales: (1) modelling of tape casting for thin ceramic layers, (2) modelling the flow of polymers in extrusion, (3) modelling the deformation process of flexible stamps...... for nanoimprint lithography, (4) modelling manufacturing of composite parts and (5) modelling the selective laser melting process. For all five examples, the emphasis is on modelling results as well as describing the models in brief mathematical details. Alongside with relevant references to the original work...

  18. Configurable multi-perspective business process models

    NARCIS (Netherlands)

    La Rosa, M.; Dumas, M.; Hofstede, ter A.H.M.; Mendling, J.

    2011-01-01

    A configurable process model provides a consolidated view of a family of business processes. It promotes the reuse of proven practices by providing analysts with a generic modeling artifact from which to derive individual process models. Unfortunately, the scope of existing notations for

  19. Probabilistic evaluation of process model matching techniques

    NARCIS (Netherlands)

    Kuss, Elena; Leopold, Henrik; van der Aa, Han; Stuckenschmidt, Heiner; Reijers, Hajo A.

    2016-01-01

    Process model matching refers to the automatic identification of corresponding activities between two process models. It represents the basis for many advanced process model analysis techniques such as the identification of similar process parts or process model search. A central problem is how to

  20. Process optimization of atomized melt deposition for the production of dispersion strengthened Al-8.5%Fe-1.2%V-1.7%Si alloys

    International Nuclear Information System (INIS)

    Hariprasad, S.; Sastry, S.M.L.; Jerina, K.L.

    1995-01-01

    Atomized melt deposition is a low cost manufacturing process with the microstructural control achieved through rapid solidification. In this process the liquid metal is disintegrated into fine droplets by gas atomization and the droplets are deposited on a substrate producing near net shape products. In the present investigation Al-8.5%Fe-1.2%V-1.7%Si alloy was produced using atomized melt deposition process to study the evolution of microstructure and assess the cooling rates and the undercooling achieved during the process. The size, morphology and the composition of second phase particles in the alloy are strong functions of the cooling rate and the undercooling and hence microstructural changes with the variation in process parameters were quantified. To define optimum conditions for the atomized melt deposition process, a mathematical model was developed. The model determines the temperature distribution of the liquid droplets during gas atomization and during the deposition stages. The model predicts the velocity distribution, cooling rates and the fraction solid, during the flight for different droplet sizes. The solidification heat transfer phenomena taking place during the atomized melt deposition process was analyzed using a finite difference method based on the enthalpy formulation

  1. Dissolution process analysis using model-free Noyes-Whitney integral equation.

    Science.gov (United States)

    Hattori, Yusuke; Haruna, Yoshimasa; Otsuka, Makoto

    2013-02-01

    Drug dissolution process of solid dosages is theoretically described by Noyes-Whitney-Nernst equation. However, the analysis of the process is demonstrated assuming some models. Normally, the model-dependent methods are idealized and require some limitations. In this study, Noyes-Whitney integral equation was proposed and applied to represent the drug dissolution profiles of a solid formulation via the non-linear least squares (NLLS) method. The integral equation is a model-free formula involving the dissolution rate constant as a parameter. In the present study, several solid formulations were prepared via changing the blending time of magnesium stearate (MgSt) with theophylline monohydrate, α-lactose monohydrate, and crystalline cellulose. The formula could excellently represent the dissolution profile, and thereby the rate constant and specific surface area could be obtained by NLLS method. Since the long time blending coated the particle surface with MgSt, it was found that the water permeation was disturbed by its layer dissociating into disintegrant particles. In the end, the solid formulations were not disintegrated; however, the specific surface area gradually increased during the process of dissolution. The X-ray CT observation supported this result and demonstrated that the rough surface was dominant as compared to dissolution, and thus, specific surface area of the solid formulation gradually increased. Copyright © 2012 Elsevier B.V. All rights reserved.

  2. Syntax highlighting in business process models

    NARCIS (Netherlands)

    Reijers, H.A.; Freytag, T.; Mendling, J.; Eckleder, A.

    2011-01-01

    Sense-making of process models is an important task in various phases of business process management initiatives. Despite this, there is currently hardly any support in business process modeling tools to adequately support model comprehension. In this paper we adapt the concept of syntax

  3. Genetic Process Mining: Alignment-based Process Model Mutation

    NARCIS (Netherlands)

    Eck, van M.L.; Buijs, J.C.A.M.; Dongen, van B.F.; Fournier, F.; Mendling, J.

    2015-01-01

    The Evolutionary Tree Miner (ETM) is a genetic process discovery algorithm that enables the user to guide the discovery process based on preferences with respect to four process model quality dimensions: replay fitness, precision, generalization and simplicity. Traditionally, the ETM algorithm uses

  4. The Mathematical Model High Energy Collisions Process Hadron-Nucleus

    International Nuclear Information System (INIS)

    Wojciechowski, A.; Strugalska-Gola, E.; Strugalski, Z.

    2002-01-01

    During the passage high energy hadron by the heavy nucleus emitted are nucleons and many other particles from which most more group are nucleons and mesons π + π - π 0 . in this work we will present the mathematical model which is a simplified description of basic processes in the interior of the nucleus during passing of the hadron by the nucleus. Result of calculations we will compare with experimental results. Experimental data are based on photographs of 180 litre xenon bubble chambers (180 1 KKP) of Institute of Theoretical and Experimental Physics in Moscow (ITEF, Moscow) irradiated with the beam of mesons π - with momentum 3.5 GeV/c. (author)

  5. Soil Methanotrophy Model (MeMo v1.0): a process-based model to quantify global uptake of atmospheric methane by soil

    Science.gov (United States)

    Murguia-Flores, Fabiola; Arndt, Sandra; Ganesan, Anita L.; Murray-Tortarolo, Guillermo; Hornibrook, Edward R. C.

    2018-06-01

    Soil bacteria known as methanotrophs are the sole biological sink for atmospheric methane (CH4), a potent greenhouse gas that is responsible for ˜ 20 % of the human-driven increase in radiative forcing since pre-industrial times. Soil methanotrophy is controlled by a plethora of factors, including temperature, soil texture, moisture and nitrogen content, resulting in spatially and temporally heterogeneous rates of soil methanotrophy. As a consequence, the exact magnitude of the global soil sink, as well as its temporal and spatial variability, remains poorly constrained. We developed a process-based model (Methanotrophy Model; MeMo v1.0) to simulate and quantify the uptake of atmospheric CH4 by soils at the global scale. MeMo builds on previous models by Ridgwell et al. (1999) and Curry (2007) by introducing several advances, including (1) a general analytical solution of the one-dimensional diffusion-reaction equation in porous media, (2) a refined representation of nitrogen inhibition on soil methanotrophy, (3) updated factors governing the influence of soil moisture and temperature on CH4 oxidation rates and (4) the ability to evaluate the impact of autochthonous soil CH4 sources on uptake of atmospheric CH4. We show that the improved structural and parametric representation of key drivers of soil methanotrophy in MeMo results in a better fit to observational data. A global simulation of soil methanotrophy for the period 1990-2009 using MeMo yielded an average annual sink of 33.5 ± 0.6 Tg CH4 yr-1. Warm and semi-arid regions (tropical deciduous forest and open shrubland) had the highest CH4 uptake rates of 602 and 518 mg CH4 m-2 yr-1, respectively. In these regions, favourable annual soil moisture content ( ˜ 20 % saturation) and low seasonal temperature variations (variations < ˜ 6 °C) provided optimal conditions for soil methanotrophy and soil-atmosphere gas exchange. In contrast to previous model analyses, but in agreement with recent observational data

  6. Functional integral and effective Hamiltonian t-J-V model of strongly correlated electron system

    International Nuclear Information System (INIS)

    Belinicher, V.I.; Chertkov, M.V.

    1990-09-01

    The functional integral representation for the generating functional of t-J-V model is obtained. In the case close to half filling this functional integral representation reduces the conventional Hamiltonian of t-J-V model to the Hamiltonian of the system containing holes and spins 1/2 at each lattice size. This effective Hamiltonian coincides with that one obtained one of the authors by different method. This Hamiltonian and its dynamical variables can be used for description of different magnetic phases of t-J-V model. (author). 16 refs

  7. Effective Connectivity between Ventral Occipito-Temporal and Ventral Inferior Frontal Cortex during Lexico-Semantic Processing. A Dynamic Causal Modeling Study

    Directory of Open Access Journals (Sweden)

    Marcela Perrone-Bertolotti

    2017-06-01

    Full Text Available It has been suggested that dorsal and ventral pathways support distinct aspects of language processing. Yet, the full extent of their involvement and their inter-regional connectivity in visual word recognition is still unknown. Studies suggest that they might reflect the dual-route model of reading, with the dorsal pathway more involved in grapho-phonological conversion during phonological tasks, and the ventral pathway performing lexico-semantic access during semantic tasks. Furthermore, this subdivision is also suggested at the level of the inferior frontal cortex, involving ventral and dorsal parts for lexico-semantic and phonological processing, respectively. In the present study, we assessed inter-regional brain connectivity and task-induced modulations of brain activity during a phoneme detection and semantic categorization tasks, using fMRI in healthy subject. We used a dynamic causal modeling approach to assess inter-regional connectivity and task demand modulation within the dorsal and ventral pathways, including the following network components: the ventral occipito-temporal cortex (vOTC; dorsal and ventral, the superior temporal gyrus (STG; dorsal, the dorsal inferior frontal gyrus (dIFG; dorsal, and the ventral IFG (vIFG; ventral. We report three distinct inter-regional interactions supporting orthographic information transfer from vOTC to other language regions (vOTC -> STG, vOTC -> vIFG and vOTC -> dIFG regardless of task demands. Moreover, we found that (a during semantic processing (direct ventral pathway the vOTC -> vIFG connection strength specifically increased and (b a lack of modulation of the vOTC -> dIFG connection strength by the task that could suggest a more general involvement of the dorsal pathway during visual word recognition. Results are discussed in terms of anatomo-functional connectivity of visual word recognition network.

  8. The triconnected abstraction of process models

    OpenAIRE

    Polyvyanyy, Artem; Smirnov, Sergey; Weske, Mathias

    2009-01-01

    Contents: Artem Polyvanny, Sergey Smirnow, and Mathias Weske The Triconnected Abstraction of Process Models 1 Introduction 2 Business Process Model Abstraction 3 Preliminaries 4 Triconnected Decomposition 4.1 Basic Approach for Process Component Discovery 4.2 SPQR-Tree Decomposition 4.3 SPQR-Tree Fragments in the Context of Process Models 5 Triconnected Abstraction 5.1 Abstraction Rules 5.2 Abstraction Algorithm 6 Related Work and Conclusions

  9. Modeling Suspension and Continuation of a Process

    Directory of Open Access Journals (Sweden)

    Oleg Svatos

    2012-04-01

    Full Text Available This work focuses on difficulties an analyst encounters when modeling suspension and continuation of a process in contemporary process modeling languages. As a basis there is introduced general lifecycle of an activity which is then compared to activity lifecycles supported by individual process modeling languages. The comparison shows that the contemporary process modeling languages cover the defined general lifecycle of an activity only partially. There are picked two popular process modeling languages and there is modeled real example, which reviews how the modeling languages can get along with their lack of native support of suspension and continuation of an activity. Upon the unsatisfying results of the contemporary process modeling languages in the modeled example, there is presented a new process modeling language which, as demonstrated, is capable of capturing suspension and continuation of an activity in much simpler and precise way.

  10. Mõtteid raamatust / Koost. Heigi Aaviste

    Index Scriptorium Estoniae

    1995-01-01

    Aut.: Maire Liivamets, Paul Luhtein, Paul-Eerik Rummo, Maksim Gorki, Johannes Semper, R.Descartes, Valmar Adams, Lilli Promet, Paul Rummo, Cicero, Honoré de Balzac, Aleksandr Pushkin, Epikuros, Debora Vaarandi, Arvi Siig, Helene Johani, Juhan Smuul, Friedebert Tuglas

  11. Perturbative extension of the standard model with a 125 GeV Higgs and Magnetic Dark Matter

    DEFF Research Database (Denmark)

    Dissauer, Karin; Frandsen, Mads Toudal; Hapola, Tuomas

    2012-01-01

    among several direct dark matter search experiments. We further constrain the parameters of the underlying theory using results from the Large Hadron Collider. The extension can accommodate the recently observed properties of the Higgs-like state and leads to interesting predictions. Finally we show......We introduce a perturbative extension of the standard model featuring a new dark matter sector together with a 125 GeV Higgs. The new sector consists of a vector-like heavy electron E, a complex scalar electron S and a standard model singlet Dirac fermion \\chi. The interactions among the dark...... matter candidate \\chi and the standard model particles occur via loop-induced processes involving the operator SE\\chi y, with y being the Yukawa-like coupling. The model is an explicit underlying realization of the light magnetic dark matter effective model introduced earlier to alleviate the tension...

  12. Strategy for design NIR calibration sets based on process spectrum and model space: An innovative approach for process analytical technology.

    Science.gov (United States)

    Cárdenas, V; Cordobés, M; Blanco, M; Alcalà, M

    2015-10-10

    The pharmaceutical industry is under stringent regulations on quality control of their products because is critical for both, productive process and consumer safety. According to the framework of "process analytical technology" (PAT), a complete understanding of the process and a stepwise monitoring of manufacturing are required. Near infrared spectroscopy (NIRS) combined with chemometrics have lately performed efficient, useful and robust for pharmaceutical analysis. One crucial step in developing effective NIRS-based methodologies is selecting an appropriate calibration set to construct models affording accurate predictions. In this work, we developed calibration models for a pharmaceutical formulation during its three manufacturing stages: blending, compaction and coating. A novel methodology is proposed for selecting the calibration set -"process spectrum"-, into which physical changes in the samples at each stage are algebraically incorporated. Also, we established a "model space" defined by Hotelling's T(2) and Q-residuals statistics for outlier identification - inside/outside the defined space - in order to select objectively the factors to be used in calibration set construction. The results obtained confirm the efficacy of the proposed methodology for stepwise pharmaceutical quality control, and the relevance of the study as a guideline for the implementation of this easy and fast methodology in the pharma industry. Copyright © 2015 Elsevier B.V. All rights reserved.

  13. Declarative modeling for process supervision

    International Nuclear Information System (INIS)

    Leyval, L.

    1989-01-01

    Our work is a contribution to computer aided supervision of continuous processes. It is inspired by an area of Artificial Intelligence: qualitative physics. Here, supervision is based on a model which continuously provides operators with a synthetic view of the process; but this model is founded on general principles of control theory rather than on physics. It involves concepts such as high gain or small time response. It helps in linking temporally the evolution of various variables. Moreover, the model provides predictions of the future behaviour of the process, which allows action advice and alarm filtering. This should greatly reduce the famous cognitive overload associated to any complex and dangerous evolution of the process

  14. The short-circuit test results of 6.9 kV/2.3 kV 400 kVA-class YBCO model transformer

    International Nuclear Information System (INIS)

    Tomioka, A.; Otonari, T.; Ogata, T.; Iwakuma, M.; Okamoto, H.; Hayashi, H.; Iijima, Y.; Saito, T.; Gosho, Y.; Tanabe, K.; Izumi, T.; Shiohara, Y.

    2011-01-01

    The 6.9 kV/2.3 kV 400 kVA-class single-phase YBCO model transformer with the YBCO tape with copper tape was manufactured for short-circuit current test. Short-circuit test was performed and the short-circuit current of primary winding was 346 A which was about six times larger than the rated current. The I-V characteristics of the winding did not change before and after the test. The transformer withstood short-circuit current. We are planning to turn the result into a consideration of a 66 kV/6.9 kV-20 MVA-class three-phase superconducting transformer. We are developing an elemental technology for 66 kV/6.9 kV 20 MVA-class power transformer with YBCO conductors. The protection of short-circuit technology is one of the elemental technologies for HTS transformer. Since short-circuit current is much higher than critical current of YBCO tape, there is a possibility that superconducting characteristics may be damaged during short-circuit period. We made a conductor to compose the YBCO tape with copper tape. We manufactured 6.9 kV/2.3 kV 400 kVA-class YBCO model transformer using this conductor and performed short-circuit current test. The short-circuit current of primary winding was 346 A which was about six times larger than the rated current. The I-V characteristics of the winding did not change before and after the test. We may consider this conductor withstands short-circuit current.

  15. ECONOMIC MODELING PROCESSES USING MATLAB

    Directory of Open Access Journals (Sweden)

    Anamaria G. MACOVEI

    2008-06-01

    Full Text Available To study economic phenomena and processes using mathem atical modeling, and to determine the approximatesolution to a problem we need to choose a method of calculation and a numerical computer program, namely thepackage of programs MatLab. Any economic process or phenomenon is a mathematical description of h is behavior,and thus draw up an economic and mathematical model that has the following stages: formulation of the problem, theanalysis process modeling, the production model and design verification, validation and implementation of the model.This article is presented an economic model and its modeling is using mathematical equations and software packageMatLab, which helps us approximation effective solution. As data entry is considered the net cost, the cost of direct andtotal cost and the link between them. I presented the basic formula for determining the total cost. Economic modelcalculations were made in MatLab software package and with graphic representation of its interpretation of the resultsachieved in terms of our specific problem.

  16. Styles in business process modeling: an exploration and a model

    NARCIS (Netherlands)

    Pinggera, J.; Soffer, P.; Fahland, D.; Weidlich, M.; Zugal, S.; Weber, B.; Reijers, H.A.; Mendling, J.

    2015-01-01

    Business process models are an important means to design, analyze, implement, and control business processes. As with every type of conceptual model, a business process model has to meet certain syntactic, semantic, and pragmatic quality requirements to be of value. For many years, such quality

  17. Composing Models of Geographic Physical Processes

    Science.gov (United States)

    Hofer, Barbara; Frank, Andrew U.

    Processes are central for geographic information science; yet geographic information systems (GIS) lack capabilities to represent process related information. A prerequisite to including processes in GIS software is a general method to describe geographic processes independently of application disciplines. This paper presents such a method, namely a process description language. The vocabulary of the process description language is derived formally from mathematical models. Physical processes in geography can be described in two equivalent languages: partial differential equations or partial difference equations, where the latter can be shown graphically and used as a method for application specialists to enter their process models. The vocabulary of the process description language comprises components for describing the general behavior of prototypical geographic physical processes. These process components can be composed by basic models of geographic physical processes, which is shown by means of an example.

  18. V&V Within Reuse-Based Software Engineering

    Science.gov (United States)

    Addy, Edward A.

    1996-01-01

    Verification and Validation (V&V) is used to increase the level of assurance of critical software, particularly that of safety-critical and mission-critical software. V&V is a systems engineering discipline that evaluates the software in a systems context, and is currently applied during the development of a specific application system. In order to bring the effectiveness of V&V to bear within reuse-based software engineering, V&V must be incorporated within the domain engineering process.

  19. Load modeling for sharp V-cutter cutting litchi ( Litchi chinensis Sonn ...

    African Journals Online (AJOL)

    harvesting. Cutting load is a key parameter for 'hand-held auto-picker' operation. However, there is still no suitable model for cutting load setting. Hence, a model describing the relationship among cutting load, blade angle and friction coefficient was developed for cutting operation by sharp V-cutters. The model was based ...

  20. The Need for V&V in Reuse-Based Software Engineering

    Science.gov (United States)

    Addy, Edward A.

    1997-01-01

    V&V is currently performed during application development for many systems, especially safety-critical and mission-critical systems. The V&V process is intended to discover errors, especially errors related to entire' domain or product line rather than a critical processing, as early as possible during the development process. The system application provides the context under which the software artifacts are validated. engineering. This paper describes a framework that extends V&V from an individual application system to a product line of systems that are developed within an architecture-based software engineering environment. This framework includes the activities of traditional application-level V&V, and extends these activities into the transition between domain engineering and application engineering. The framework includes descriptions of the types of activities to be performed during each of the life-cycle phases, and provides motivation for activities.

  1. An empirical-statistical model for laser cladding of Ti-6Al-4V powder on Ti-6Al-4V substrate

    Science.gov (United States)

    Nabhani, Mohammad; Razavi, Reza Shoja; Barekat, Masoud

    2018-03-01

    In this article, Ti-6Al-4V powder alloy was directly deposited on Ti-6Al-4V substrate using laser cladding process. In this process, some key parameters such as laser power (P), laser scanning rate (V) and powder feeding rate (F) play important roles. Using linear regression analysis, this paper develops the empirical-statistical relation between these key parameters and geometrical characteristics of single clad tracks (i.e. clad height, clad width, penetration depth, wetting angle, and dilution) as a combined parameter (PαVβFγ). The results indicated that the clad width linearly depended on PV-1/3 and powder feeding rate had no effect on it. The dilution controlled by a combined parameter as VF-1/2 and laser power was a dispensable factor. However, laser power was the dominant factor for the clad height, penetration depth, and wetting angle so that they were proportional to PV-1F1/4, PVF-1/8, and P3/4V-1F-1/4, respectively. Based on the results of correlation coefficient (R > 0.9) and analysis of residuals, it was confirmed that these empirical-statistical relations were in good agreement with the measured values of single clad tracks. Finally, these relations led to the design of a processing map that can predict the geometrical characteristics of the single clad tracks based on the key parameters.

  2. The Structured Process Modeling Theory (SPMT) : a cognitive view on why and how modelers benefit from structuring the process of process modeling

    NARCIS (Netherlands)

    Claes, J.; Vanderfeesten, I.T.P.; Gailly, F.; Grefen, P.W.P.J.; Poels, G.

    2015-01-01

    After observing various inexperienced modelers constructing a business process model based on the same textual case description, it was noted that great differences existed in the quality of the produced models. The impression arose that certain quality issues originated from cognitive failures

  3. Background model systematics for the Fermi GeV excess

    Energy Technology Data Exchange (ETDEWEB)

    Calore, Francesca; Cholis, Ilias; Weniger, Christoph

    2015-03-01

    The possible gamma-ray excess in the inner Galaxy and the Galactic center (GC) suggested by Fermi-LAT observations has triggered a large number of studies. It has been interpreted as a variety of different phenomena such as a signal from WIMP dark matter annihilation, gamma-ray emission from a population of millisecond pulsars, or emission from cosmic rays injected in a sequence of burst-like events or continuously at the GC. We present the first comprehensive study of model systematics coming from the Galactic diffuse emission in the inner part of our Galaxy and their impact on the inferred properties of the excess emission at Galactic latitudes 2° < |b| < 20° and 300 MeV to 500 GeV. We study both theoretical and empirical model systematics, which we deduce from a large range of Galactic diffuse emission models and a principal component analysis of residuals in numerous test regions along the Galactic plane. We show that the hypothesis of an extended spherical excess emission with a uniform energy spectrum is compatible with the Fermi-LAT data in our region of interest at 95% CL. Assuming that this excess is the extended counterpart of the one seen in the inner few degrees of the Galaxy, we derive a lower limit of 10.0° (95% CL) on its extension away from the GC. We show that, in light of the large correlated uncertainties that affect the subtraction of the Galactic diffuse emission in the relevant regions, the energy spectrum of the excess is equally compatible with both a simple broken power-law of break energy E(break) = 2.1 ± 0.2 GeV, and with spectra predicted by the self-annihilation of dark matter, implying in the case of bar bb final states a dark matter mass of m(χ)=49(+6.4)(-)(5.4)  GeV.

  4. Integrated modelling in materials and process technology

    DEFF Research Database (Denmark)

    Hattel, Jesper Henri

    2008-01-01

    Integrated modelling of entire process sequences and the subsequent in-service conditions, and multiphysics modelling of the single process steps are areas that increasingly support optimisation of manufactured parts. In the present paper, three different examples of modelling manufacturing...... processes from the viewpoint of combined materials and process modelling are presented: solidification of thin walled ductile cast iron, integrated modelling of spray forming and multiphysics modelling of friction stir welding. The fourth example describes integrated modelling applied to a failure analysis...

  5. The impact of working memory and the “process of process modelling” on model quality: Investigating experienced versus inexperienced modellers

    Science.gov (United States)

    Martini, Markus; Pinggera, Jakob; Neurauter, Manuel; Sachse, Pierre; Furtner, Marco R.; Weber, Barbara

    2016-01-01

    A process model (PM) represents the graphical depiction of a business process, for instance, the entire process from online ordering a book until the parcel is delivered to the customer. Knowledge about relevant factors for creating PMs of high quality is lacking. The present study investigated the role of cognitive processes as well as modelling processes in creating a PM in experienced and inexperienced modellers. Specifically, two working memory (WM) functions (holding and processing of information and relational integration) and three process of process modelling phases (comprehension, modelling, and reconciliation) were related to PM quality. Our results show that the WM function of relational integration was positively related to PM quality in both modelling groups. The ratio of comprehension phases was negatively related to PM quality in inexperienced modellers and the ratio of reconciliation phases was positively related to PM quality in experienced modellers. Our research reveals central cognitive mechanisms in process modelling and has potential practical implications for the development of modelling software and teaching the craft of process modelling. PMID:27157858

  6. Detecting Difference between Process Models Based on the Refined Process Structure Tree

    Directory of Open Access Journals (Sweden)

    Jing Fan

    2017-01-01

    Full Text Available The development of mobile workflow management systems (mWfMS leads to large number of business process models. In the meantime, the location restriction embedded in mWfMS may result in different process models for a single business process. In order to help users quickly locate the difference and rebuild the process model, detecting the difference between different process models is needed. Existing detection methods either provide a dissimilarity value to represent the difference or use predefined difference template to generate the result, which cannot reflect the entire composition of the difference. Hence, in this paper, we present a new approach to solve this problem. Firstly, we parse the process models to their corresponding refined process structure trees (PSTs, that is, decomposing a process model into a hierarchy of subprocess models. Then we design a method to convert the PST to its corresponding task based process structure tree (TPST. As a consequence, the problem of detecting difference between two process models is transformed to detect difference between their corresponding TPSTs. Finally, we obtain the difference between two TPSTs based on the divide and conquer strategy, where the difference is described by an edit script and we make the cost of the edit script close to minimum. The extensive experimental evaluation shows that our method can meet the real requirements in terms of precision and efficiency.

  7. Strategies to Automatically Derive a Process Model from a Configurable Process Model Based on Event Data

    Directory of Open Access Journals (Sweden)

    Mauricio Arriagada-Benítez

    2017-10-01

    Full Text Available Configurable process models are frequently used to represent business workflows and other discrete event systems among different branches of large organizations: they unify commonalities shared by all branches and describe their differences, at the same time. The configuration of such models is usually done manually, which is challenging. On the one hand, when the number of configurable nodes in the configurable process model grows, the size of the search space increases exponentially. On the other hand, the person performing the configuration may lack the holistic perspective to make the right choice for all configurable nodes at the same time, since choices influence each other. Nowadays, information systems that support the execution of business processes create event data reflecting how processes are performed. In this article, we propose three strategies (based on exhaustive search, genetic algorithms and a greedy heuristic that use event data to automatically derive a process model from a configurable process model that better represents the characteristics of the process in a specific branch. These strategies have been implemented in our proposed framework and tested in both business-like event logs as recorded in a higher educational enterprise resource planning system and a real case scenario involving a set of Dutch municipalities.

  8. Fluence to absorbed foetal dose conversion coefficients for photons in 50 keV-10 GeV calculated using RPI-P models

    International Nuclear Information System (INIS)

    Taranenko, V.; Xu, X.G.

    2008-01-01

    Radiation protection of pregnant females and the foetus against ionising radiation is of particular importance to radiation protection due to high foetal radiosensitivity. The only available set of foetal conversion coefficients for photons is based on stylised models of simplified anatomy. Using the RPI-P series of pregnant female and foetus models representing 3-, 6- and 9-month gestation, a set of new fluence to absorbed foetal dose conversion coefficients has been calculated. The RPI-P anatomical models were developed using novel 3D geometry modelling techniques. Organ masses were adjusted to agree within 1% with the ICRP reference data for a pregnant female. Monte Carlo dose calculations were carried out using the MCNPX and Penelope codes for external 50 keV-10 GeV photon beams of six standard configurations. The models were voxelised at 3-mm voxel resolution. Conversion coefficients were tabulated for the three gestational periods for the whole foetus and brain. Comparison with previously published data showed deviations up to 120% for the foetal doses at 50 keV. The discrepancy can be primarily ascribed to anatomical differences. Comparison with published data for five major mother organs is also provided for the 3-month model. Since the RPI-P models exhibit a high degree of anatomical realism, the reported dataset is recommended as a reference for radiation protection of the foetus against external photon exposure. (authors)

  9. A multi-component evaporation model for beam melting processes

    Science.gov (United States)

    Klassen, Alexander; Forster, Vera E.; Körner, Carolin

    2017-02-01

    In additive manufacturing using laser or electron beam melting technologies, evaporation losses and changes in chemical composition are known issues when processing alloys with volatile elements. In this paper, a recently described numerical model based on a two-dimensional free surface lattice Boltzmann method is further developed to incorporate the effects of multi-component evaporation. The model takes into account the local melt pool composition during heating and fusion of metal powder. For validation, the titanium alloy Ti-6Al-4V is melted by selective electron beam melting and analysed using mass loss measurements and high-resolution microprobe imaging. Numerically determined evaporation losses and spatial distributions of aluminium compare well with experimental data. Predictions of the melt pool formation in bulk samples provide insight into the competition between the loss of volatile alloying elements from the irradiated surface and their advective redistribution within the molten region.

  10. Repairing process models to reflect reality

    NARCIS (Netherlands)

    Fahland, D.; Aalst, van der W.M.P.; Barros, A.; Gal, A.; Kindler, E.

    2012-01-01

    Process mining techniques relate observed behavior (i.e., event logs) to modeled behavior (e.g., a BPMN model or a Petri net). Processes models can be discovered from event logs and conformance checking techniques can be used to detect and diagnose differences between observed and modeled behavior.

  11. Business process modeling for processing classified documents using RFID technology

    Directory of Open Access Journals (Sweden)

    Koszela Jarosław

    2016-01-01

    Full Text Available The article outlines the application of the processing approach to the functional description of the designed IT system supporting the operations of the secret office, which processes classified documents. The article describes the application of the method of incremental modeling of business processes according to the BPMN model to the description of the processes currently implemented (“as is” in a manual manner and target processes (“to be”, using the RFID technology for the purpose of their automation. Additionally, the examples of applying the method of structural and dynamic analysis of the processes (process simulation to verify their correctness and efficiency were presented. The extension of the process analysis method is a possibility of applying the warehouse of processes and process mining methods.

  12. Intranuclear cascade evaporation model predictions of double differential A(p,xn) neutron cross sections and comparison with experiments at 318 MeV and 800 MeV proton energy

    International Nuclear Information System (INIS)

    Cloth, P.; Dragovitsch, P.; Filges, D.; Reul, C.

    1989-08-01

    The intranuclear-cascade evaporation model as implemented in the high energy radiation transport code HETC, subsystem of HERMES is used in the calculation of double differential cross sections of proton induced neutron production. The investigations were done on target elements C, Al, Ta, Ni, W, Pb, and U at 318 MeV incident proton energy and on C, Al, Pb, and U at 800 MeV, respectively. The predictions of the INCE model were compared with experimental data for double differential cross sections taken at 7.5 and 30 degrees scattering angles at the Los Alamos WNR facility utilizing the Time of Flight technique at LANL. The calculations performed here are part of a experimental-theoretical program within the LANL-KFA collaboration concerning medium energy cross section measurements mainly neutrons and state of the art computer code validations of these measurements. In general, the model predictions reproduce the correct neutron production for evaporation neutrons and are also in good agreement with the experimental data at high neutron energies. In the energy range dominated by preequilibrium processes an underestimation of experimental yields has to be remarked. (orig.)

  13. Mathematical modeling of a V-stack piezoelectric aileron actuation

    Directory of Open Access Journals (Sweden)

    Ioan URSU

    2016-12-01

    Full Text Available The article presents a mathematical modeling of aileron actuation that uses piezo V-shaped stacks. The aim of the actuation is the increasing of flutter speed in the context of a control law, in order to widen the flight envelope. In this way the main advantage of such a piezo actuator, the bandwidth is exploited. The mathematical model is obtained based on free body diagrams, and the numerical simulations allow a preliminary sizing of the actuator.

  14. Business Process Modelling for Measuring Quality

    NARCIS (Netherlands)

    Heidari, F.; Loucopoulos, P.; Brazier, F.M.

    2013-01-01

    Business process modelling languages facilitate presentation, communication and analysis of business processes with different stakeholders. This paper proposes an approach that drives specification and measurement of quality requirements and in doing so relies on business process models as

  15. Investigation of Mediational Processes Using Parallel Process Latent Growth Curve Modeling

    Science.gov (United States)

    Cheong, JeeWon; MacKinnon, David P.; Khoo, Siek Toon

    2010-01-01

    This study investigated a method to evaluate mediational processes using latent growth curve modeling. The mediator and the outcome measured across multiple time points were viewed as 2 separate parallel processes. The mediational process was defined as the independent variable influencing the growth of the mediator, which, in turn, affected the growth of the outcome. To illustrate modeling procedures, empirical data from a longitudinal drug prevention program, Adolescents Training and Learning to Avoid Steroids, were used. The program effects on the growth of the mediator and the growth of the outcome were examined first in a 2-group structural equation model. The mediational process was then modeled and tested in a parallel process latent growth curve model by relating the prevention program condition, the growth rate factor of the mediator, and the growth rate factor of the outcome. PMID:20157639

  16. Roll-to-Roll Processing of Inverted Polymer Solar Cells using Hydrated Vanadium(V)Oxide as a PEDOT:PSS Replacement

    DEFF Research Database (Denmark)

    Martinez, Nieves Espinosa; Dam, Henrik Friis; Tanenbaum, David M.

    2011-01-01

    roll-to-roll (R2R) processing of all layers. The devices were prepared on flexible polyethyleneterphthalate (PET) and had the structure PET/ITO/ZnO/P3HT:PCBM/V2O5·(H2O)n/Ag. The ITO and silver electrodes were processed and patterned by use of screen printing. The zinc oxide, P3HT:PCBM and vanadium(V)oxide......The use of hydrated vanadium(V)oxide as a replacement of the commonly employed hole transporting material PEDOT:PSS was explored in this work. Polymer solar cells were prepared by spin coating on glass. Polymer solar cells and modules comprising 16 serially connected cells were prepared using full...... layers were processed by slot-die coating. The hydrated vanadium(V)oxide layer was slot-die coated using an isopropanol solution of vanadyl-triisopropoxide (VTIP). Coating experiments were carried out to establish the critical thickness of the hydrated vanadium(V)oxide layer by varying the concentration...

  17. Early auditory processing in area V5/MT+ of the congenitally blind brain.

    Science.gov (United States)

    Watkins, Kate E; Shakespeare, Timothy J; O'Donoghue, M Clare; Alexander, Iona; Ragge, Nicola; Cowey, Alan; Bridge, Holly

    2013-11-13

    Previous imaging studies of congenital blindness have studied individuals with heterogeneous causes of blindness, which may influence the nature and extent of cross-modal plasticity. Here, we scanned a homogeneous group of blind people with bilateral congenital anophthalmia, a condition in which both eyes fail to develop, and, as a result, the visual pathway is not stimulated by either light or retinal waves. This model of congenital blindness presents an opportunity to investigate the effects of very early visual deafferentation on the functional organization of the brain. In anophthalmic animals, the occipital cortex receives direct subcortical auditory input. We hypothesized that this pattern of subcortical reorganization ought to result in a topographic mapping of auditory frequency information in the occipital cortex of anophthalmic people. Using functional MRI, we examined auditory-evoked activity to pure tones of high, medium, and low frequencies. Activity in the superior temporal cortex was significantly reduced in anophthalmic compared with sighted participants. In the occipital cortex, a region corresponding to the cytoarchitectural area V5/MT+ was activated in the anophthalmic participants but not in sighted controls. Whereas previous studies in the blind indicate that this cortical area is activated to auditory motion, our data show it is also active for trains of pure tone stimuli and in some anophthalmic participants shows a topographic mapping (tonotopy). Therefore, this region appears to be performing early sensory processing, possibly served by direct subcortical input from the pulvinar to V5/MT+.

  18. Development and fabrication process for ZnO based varistors for medium voltage arresters 13000 V to 34000 V; Desarrollo y produccion de varistores de ZnO dopados para media tension 13000 V a 34000 V

    Energy Technology Data Exchange (ETDEWEB)

    Gonzalez Rolon, B.; Ireta Moreno, F. [Facultad de Ingenieria Mecanica, Electrica y Electronica, Universidad de Guanajuato, Salamanca, Guanajuato (Mexico)]. E-mail: barbara@salamanca.ugto.mx; fireta@salamanca.ugto.mx

    2011-04-15

    The fabrication process for ZnO doped varistor blocks with Sb2O3/ Bi2O3 to medium tension of 13000 V to 34000 V, and was optimised in terms of a starting composition and firing temperatures with amounts for Sb2O3/ Bi2O3 ratios of 1.7 by dry pressing, and fired in the temperature 1150 degrees Celsius {+-}2 degrees Celsius. The microstructure characteristics were analysed across the surface of the varistor blocks in order to evaluate their microstructure homogeneity. Their current-voltage I-V and energy handling capabilities were also determined. Characteristics and electrical properties of the varistor blocks are discussed. [Spanish] El proceso de fabricacion para varistores de media tension de 13000 V a 34000 V de ZnO consistio en dopar con Sb2O3, Bi2O3 y optimizarlos en terminos de la composicion de inicio, con una relacion de Sb2O3/Bi2O3 de 1.7, y el uso de aditivos para ceramicas, la temperatura de sinterizado fue de 1150 grados centigrados {+-}2 grados centigrados. Se investigo la homogeneidad de la microestructura. Se determinaron las propiedades electricas mediante la determinacion de la grafica I-V. Las caracteristicas electricas son discutidas.

  19. The Adsorption Langmuir Model of Transfer Metal Ti, V and Mn on System Water-Sediment in Along Side Code River, Yogyakarta

    International Nuclear Information System (INIS)

    Rini Jati Wardani; Muzakky; Agus Taftazani

    2007-01-01

    The adsorption langmuir model of transfer metal Ti, V and Mn on system water-sediment in along side Code river, Yogyakarta has been studied. For that purpose, the study is to make prediction about adsorption langmuir model of identified metal Ti, V and Mn from upstream until downstream samples water and sediment in along side Code river. The factor influenced of langmuir adsorption on transfer metal Ti, V and Mn in system water-sediment is Total Suspended Solid (TSS). The analysis showed that between TSS with metal concentration in sediment have linear correlation. The result of calculation from curve of langmuir isotherm, showed for Ti has R 2 = 0.8006 with capacities of adsorption = 0.5 mol/l and energy of adsorption = 13.286 J/mol, V has R 2 = 0.9883 with capacities of adsorption = 0.0137 mol/l and energy of adsorption = 16.64 J/mol, Mn has R 2 = 0.9624 with capacities of adsorption 0.152 mol/l and energy of adsorption = 10.51 J/mol. The conclusion from this topic about adsorption langmuir for metal Ti, V and Mn according to energy of langmuir adsorption by chemisorption process above 10 J/mo. (author)

  20. Modeling human auditory evoked brainstem responses based on nonlinear cochlear processing

    DEFF Research Database (Denmark)

    Harte, James; Rønne, Filip Munch; Dau, Torsten

    2010-01-01

    . To generate AEPs recorded at remote locations, a convolution was made on an empirically obtained elementary unit waveform with the instantaneous discharge rate function for the corresponding AN unit. AEPs to click-trains, as well as to tone pulses at various frequencies, were both modelled and recorded...... at different stimulation levels and repetition rates. The observed nonlinearities in the recorded potential patterns, with respect to ABR wave V latencies and amplitudes, could be largely accounted for by level-dependent BM processing as well as effects of short-term neural adaptation. The present study...

  1. Soil Methanotrophy Model (MeMo v1.0: a process-based model to quantify global uptake of atmospheric methane by soil

    Directory of Open Access Journals (Sweden)

    F. Murguia-Flores

    2018-06-01

    Full Text Available Soil bacteria known as methanotrophs are the sole biological sink for atmospheric methane (CH4, a potent greenhouse gas that is responsible for  ∼  20 % of the human-driven increase in radiative forcing since pre-industrial times. Soil methanotrophy is controlled by a plethora of factors, including temperature, soil texture, moisture and nitrogen content, resulting in spatially and temporally heterogeneous rates of soil methanotrophy. As a consequence, the exact magnitude of the global soil sink, as well as its temporal and spatial variability, remains poorly constrained. We developed a process-based model (Methanotrophy Model; MeMo v1.0 to simulate and quantify the uptake of atmospheric CH4 by soils at the global scale. MeMo builds on previous models by Ridgwell et al. (1999 and Curry (2007 by introducing several advances, including (1 a general analytical solution of the one-dimensional diffusion–reaction equation in porous media, (2 a refined representation of nitrogen inhibition on soil methanotrophy, (3 updated factors governing the influence of soil moisture and temperature on CH4 oxidation rates and (4 the ability to evaluate the impact of autochthonous soil CH4 sources on uptake of atmospheric CH4. We show that the improved structural and parametric representation of key drivers of soil methanotrophy in MeMo results in a better fit to observational data. A global simulation of soil methanotrophy for the period 1990–2009 using MeMo yielded an average annual sink of 33.5 ± 0.6 Tg CH4 yr−1. Warm and semi-arid regions (tropical deciduous forest and open shrubland had the highest CH4 uptake rates of 602 and 518 mg CH4 m−2 yr−1, respectively. In these regions, favourable annual soil moisture content ( ∼  20 % saturation and low seasonal temperature variations (variations  <   ∼  6 °C provided optimal conditions for soil methanotrophy and soil–atmosphere gas exchange

  2. Selective Laser Melting Produced Ti-6Al-4V: Post-Process Heat Treatments to Achieve Superior Tensile Properties.

    Science.gov (United States)

    Ter Haar, Gerrit M; Becker, Thorsten H

    2018-01-17

    Current post-process heat treatments applied to selective laser melting produced Ti-6Al-4V do not achieve the same microstructure and therefore superior tensile behaviour of thermomechanical processed wrought Ti-6Al-4V. Due to the growing demand for selective laser melting produced parts in industry, research and development towards improved mechanical properties is ongoing. This study is aimed at developing post-process annealing strategies to improve tensile behaviour of selective laser melting produced Ti-6Al-4V parts. Optical and electron microscopy was used to study α grain morphology as a function of annealing temperature, hold time and cooling rate. Quasi-static uniaxial tensile tests were used to measure tensile behaviour of different annealed parts. It was found that elongated α'/α grains can be fragmented into equiaxial grains through applying a high temperature annealing strategy. It is shown that bi-modal microstructures achieve a superior tensile ductility to current heat treated selective laser melting produced Ti-6Al-4V samples.

  3. Behavioral conformance of artifact-centric process models

    NARCIS (Netherlands)

    Fahland, D.; Leoni, de M.; Dongen, van B.F.; Aalst, van der W.M.P.; Abramowicz, W.

    2011-01-01

    The use of process models in business information systems for analysis, execution, and improvement of processes assumes that the models describe reality. Conformance checking is a technique to validate how good a given process model describes recorded executions of the actual process. Recently,

  4. Process Parameters Optimization of 14nm MOSFET Using 2-D Analytical Modelling

    Directory of Open Access Journals (Sweden)

    Noor Faizah Z.A.

    2016-01-01

    Full Text Available This paper presents the modeling and optimization of 14nm gate length CMOS transistor which is down-scaled from previous 32nm gate length. High-k metal gate material was used in this research utilizing Hafnium Dioxide (HfO2 as dielectric and Tungsten Silicide (WSi2 and Titanium Silicide (TiSi2 as a metal gate for NMOS and PMOS respectively. The devices are fabricated virtually using ATHENA module and characterized its performance evaluation via ATLAS module; both in Virtual Wafer Fabrication (VWF of Silvaco TCAD Tools. The devices were then optimized through a process parameters variability using L9 Taguchi Method. There were four process parameter with two noise factor of different values were used to analyze the factor effect. The results show that the optimal value for both transistors are well within ITRS 2013 prediction where VTH and IOFF are 0.236737V and 6.995705nA/um for NMOS device and 0.248635 V and 5.26nA/um for PMOS device respectively.

  5. KENO3D visualization tool for KENO V.a geometry models

    International Nuclear Information System (INIS)

    Bowman, S.M.; Horwedel, J.E.

    1999-01-01

    The standardized computer analyses for licensing evaluations (SCALE) computer software system developed at Oak Ridge National Laboratory (ORNL) is widely used and accepted around the world for criticality safety analyses. SCALE includes the well-known KENO V.a three-dimensional Monte Carlo criticality computer code. Criticality safety analysis often require detailed modeling of complex geometries. Checking the accuracy of these models can be enhanced by effective visualization tools. To address this need, ORNL has recently developed a powerful state-of-the-art visualization tool called KENO3D that enables KENO V.a users to interactively display their three-dimensional geometry models. The interactive options include the following: (1) having shaded or wireframe images; (2) showing standard views, such as top view, side view, front view, and isometric three-dimensional view; (3) rotating the model; (4) zooming in on selected locations; (5) selecting parts of the model to display; (6) editing colors and displaying legends; (7) displaying properties of any unit in the model; (8) creating cutaway views; (9) removing units from the model; and (10) printing image or saving image to common graphics formats

  6. Development and Evaluation of the Biogenic Emissions Inventory System (BEIS) Model v3.6

    Science.gov (United States)

    We have developed new canopy emission algorithms and land use data for BEIS v3.6. Simulations with BEIS v3.4 and BEIS v3.6 in CMAQ v5.0.2 are compared these changes to the Model of Emissions of Gases and Aerosols from Nature (MEGAN) and evaluated the simulations against observati...

  7. A flavour independent search for the Higgsstrahlung process in e+e- collisions at centre-of-mass energies from 189 to 209 GeV

    CERN Document Server

    Pascolo, J M

    2001-01-01

    Higgs boson decays to b or tau pairs, dominant in the standard model, can be suppressed in extended models. A flavour-independent search for the Higgsstrahlung process has been carried out to treat all possible hadronic decays of the Higgs boson on an equal footing. The analysis is based on the data collected by the ALEPH detector at LEP at centre-of-mass energies from 189 to 209 GeV. A 95\\% CL lower mass limit of is obtained for a cross-section equal to that of the standard model and a 100\\% hadronic branching fraction.

  8. Cost Models for MMC Manufacturing Processes

    Science.gov (United States)

    Elzey, Dana M.; Wadley, Haydn N. G.

    1996-01-01

    Processes for the manufacture of advanced metal matrix composites are rapidly approaching maturity in the research laboratory and there is growing interest in their transition to industrial production. However, research conducted to date has almost exclusively focused on overcoming the technical barriers to producing high-quality material and little attention has been given to the economical feasibility of these laboratory approaches and process cost issues. A quantitative cost modeling (QCM) approach was developed to address these issues. QCM are cost analysis tools based on predictive process models relating process conditions to the attributes of the final product. An important attribute, of the QCM approach is the ability to predict the sensitivity of material production costs to product quality and to quantitatively explore trade-offs between cost and quality. Applications of the cost models allow more efficient direction of future MMC process technology development and a more accurate assessment of MMC market potential. Cost models were developed for two state-of-the art metal matrix composite (MMC) manufacturing processes: tape casting and plasma spray deposition. Quality and Cost models are presented for both processes and the resulting predicted quality-cost curves are presented and discussed.

  9. Conceptual modelling of human resource evaluation process

    Directory of Open Access Journals (Sweden)

    Negoiţă Doina Olivia

    2017-01-01

    Full Text Available Taking into account the highly diverse tasks which employees have to fulfil due to complex requirements of nowadays consumers, the human resource within an enterprise has become a strategic element for developing and exploiting products which meet the market expectations. Therefore, organizations encounter difficulties when approaching the human resource evaluation process. Hence, the aim of the current paper is to design a conceptual model of the aforementioned process, which allows the enterprises to develop a specific methodology. In order to design the conceptual model, Business Process Modelling instruments were employed - Adonis Community Edition Business Process Management Toolkit using the ADONIS BPMS Notation. The conceptual model was developed based on an in-depth secondary research regarding the human resource evaluation process. The proposed conceptual model represents a generic workflow (sequential and/ or simultaneously activities, which can be extended considering the enterprise’s needs regarding their requirements when conducting a human resource evaluation process. Enterprises can benefit from using software instruments for business process modelling as they enable process analysis and evaluation (predefined / specific queries and also model optimization (simulations.

  10. Product- and Process Units in the CRITT Translation Process Research Database

    DEFF Research Database (Denmark)

    Carl, Michael

    than 300 hours of text production. The database provides the raw logging data, as well as Tables of pre-processed product- and processing units. The TPR-DB includes various types of simple and composed product and process units that are intended to support the analysis and modelling of human text......The first version of the "Translation Process Research Database" (TPR DB v1.0) was released In August 2012, containing logging data of more than 400 translation and text production sessions. The current version of the TPR DB, (v1.4), contains data from more than 940 sessions, which represents more...

  11. Process analysis and modeling of a single-step lutein extraction method for wet microalgae.

    Science.gov (United States)

    Gong, Mengyue; Wang, Yuruihan; Bassi, Amarjeet

    2017-11-01

    Lutein is a commercial carotenoid with potential health benefits. Microalgae are alternative sources for the lutein production in comparison to conventional approaches using marigold flowers. In this study, a process analysis of a single-step simultaneous extraction, saponification, and primary purification process for free lutein production from wet microalgae biomass was carried out. The feasibility of binary solvent mixtures for wet biomass extraction was successfully demonstrated, and the extraction kinetics of lutein from chloroplast in microalgae were first evaluated. The effects of types of organic solvent, solvent polarity, cell disruption method, and alkali and solvent usage on lutein yields were examined. A mathematical model based on Fick's second law of diffusion was applied to model the experimental data. The mass transfer coefficients were used to estimate the extraction rates. The extraction rate was found more significantly related with alkali ratio to solvent than to biomass. The best conditions for extraction efficiency were found to be pre-treatment with ultrasonication at 0.5 s working cycle per second, react 0.5 h in 0.27 L/g solvent to biomass ratio, and 1:3 ether/ethanol (v/v) with 1.25 g KOH/L. The entire process can be controlled within 1 h and yield over 8 mg/g lutein, which is more economical for scale-up.

  12. Impaired spatial processing in a mouse model of fragile X syndrome.

    Science.gov (United States)

    Ghilan, Mohamed; Bettio, Luis E B; Noonan, Athena; Brocardo, Patricia S; Gil-Mohapel, Joana; Christie, Brian R

    2018-05-17

    Fragile X syndrome (FXS) is the most common form of inherited intellectual impairment. The Fmr1 -/y mouse model has been previously shown to have deficits in context discrimination tasks but not in the elevated plus-maze. To further characterize this FXS mouse model and determine whether hippocampal-mediated behaviours are affected in these mice, dentate gyrus (DG)-dependent spatial processing and Cornu ammonis 1 (CA1)-dependent temporal order discrimination tasks were evaluated. In agreement with previous findings of long-term potentiation deficits in the DG of this transgenic model of FXS, the results reported here demonstrate that Fmr1 -/y mice perform poorly in the DG-dependent metric change spatial processing task. However, Fmr1 -/y mice did not present deficits in the CA1-dependent temporal order discrimination task, and were able to remember the order in which objects were presented to them to the same extent as their wild-type littermate controls. These data suggest that the previously reported subregional-specific differences in hippocampal synaptic plasticity observed in the Fmr1 -/y mouse model may manifest as selective behavioural deficits in hippocampal-dependent tasks. Crown Copyright © 2018. Published by Elsevier B.V. All rights reserved.

  13. Modeling Dynamic Systems with Efficient Ensembles of Process-Based Models.

    Directory of Open Access Journals (Sweden)

    Nikola Simidjievski

    Full Text Available Ensembles are a well established machine learning paradigm, leading to accurate and robust models, predominantly applied to predictive modeling tasks. Ensemble models comprise a finite set of diverse predictive models whose combined output is expected to yield an improved predictive performance as compared to an individual model. In this paper, we propose a new method for learning ensembles of process-based models of dynamic systems. The process-based modeling paradigm employs domain-specific knowledge to automatically learn models of dynamic systems from time-series observational data. Previous work has shown that ensembles based on sampling observational data (i.e., bagging and boosting, significantly improve predictive performance of process-based models. However, this improvement comes at the cost of a substantial increase of the computational time needed for learning. To address this problem, the paper proposes a method that aims at efficiently learning ensembles of process-based models, while maintaining their accurate long-term predictive performance. This is achieved by constructing ensembles with sampling domain-specific knowledge instead of sampling data. We apply the proposed method to and evaluate its performance on a set of problems of automated predictive modeling in three lake ecosystems using a library of process-based knowledge for modeling population dynamics. The experimental results identify the optimal design decisions regarding the learning algorithm. The results also show that the proposed ensembles yield significantly more accurate predictions of population dynamics as compared to individual process-based models. Finally, while their predictive performance is comparable to the one of ensembles obtained with the state-of-the-art methods of bagging and boosting, they are substantially more efficient.

  14. Synthesis of (Cr,V){sub 2}(C,N) solid solution powders by thermal processing precursors

    Energy Technology Data Exchange (ETDEWEB)

    Liu, Anrui [School of Materials Science & Engineering, Sichuan University, Chengdu, 610065 (China); Liu, Ying [School of Materials Science & Engineering, Sichuan University, Chengdu, 610065 (China); Key Laboratory of Advanced Special Material & Technology, Ministry of Education, Chengdu, 610065 (China); Ma, Shiqing; Qiu, Yuchong; Rong, Pengcheng; Ye, Jinwen [School of Materials Science & Engineering, Sichuan University, Chengdu, 610065 (China)

    2017-06-01

    The single-phase (Cr,V){sub 2}(C,N) solid solution powders were fabricated via carbothermal reduction-nitridation (CRN) processing technique. The effects of heat treatment temperature, nitrogen pressure and carbon proportion were experimentally studied in detail by X-ray Diffraction (XRD), Scanning Electron Microscope (SEM), Transmission Electron Microscope (TEM) and thermal analysis. The chemical transformations of vanadium and chromium compounds were as follows: precursors → V{sub 2}O{sub 3}, Cr{sub 2}O{sub 3} → Cr{sub 3}C{sub 2}, Cr{sub 2}O{sub 3}, (Cr,V){sub 2}(C,N) → (Cr,V){sub 2}(C,N). When the heat-treated temperature was below 1200 °C, chromium oxides didn’t completely react. However, higher temperature ∼1300 °C could not only lead to the segregation of some nitrides and carbon black, but also to the occurrence of fiber-bridged particles. The system nitrogen pressure over 0.03 MPa would cause a subtle transformation of (Cr,V){sub 2}(C,N) to VCrN{sub 2}. When the carbon proportion was below 15 wt%, the oxides could not be completely reduced, while when the carbon proportion was above 15.5 wt%, some undesired carbides, like Cr{sub 23}C{sub 6} and Cr{sub 3}C{sub 2}, would form. Ultimately, the homogeneously distributed pure-phase (Cr,V){sub 2}(C,N) spherical particles with the average size of ∼1.5 μm were obtained at the optimal conditions of the treatment of precursors at 1200 °C for 1 h with the nitrogen pressure of 0.03 MPa and carbon content of 15.5 wt%. The chemical composition of the solid solution with the optimal process could be drawn as (Cr{sub 0.85}V{sub 0.15}){sub 2}(C{sub 0.57}N{sub 0.43}). Thermal processing precursors method shows the advantages of lower synthesis temperature, shorter period and finer particles when comparing with the conventional preparations. - Highlights: • Single phase of (Cr,V){sub 2}(C,N) powders were synthesized for the first time. • Precursors were used to prepared the powders by carbothermal

  15. Polymer-bound oxidovanadium(IV) and dioxidovanadium(V) complexes as catalysts for the oxidative desulfurization of model fuel diesel.

    Science.gov (United States)

    Maurya, Mannar R; Arya, Aarti; Kumar, Amit; Kuznetsov, Maxim L; Avecilla, Fernando; Costa Pessoa, João

    2010-07-19

    The Schiff base (Hfsal-dmen) derived from 3-formylsalicylic acid and N,N-dimethyl ethylenediamine has been covalently bonded to chloromethylated polystyrene to give the polymer-bound ligand, PS-Hfsal-dmen (I). Treatment of PS-Hfsal-dmen with [V(IV)O(acac)(2)] in the presence of MeOH gave the oxidovanadium(IV) complex PS-[V(IV)O(fsal-dmen)(MeO)] (1). On aerial oxidation in methanol, complex 1 was oxidized to PS-[V(V)O(2)(fsal-dmen)] (2). The corresponding neat complexes, [V(IV)O(sal-dmen)(acac)] (3) and [V(V)O(2)(sal-dmen)] (4) were similarly prepared. All these complexes are characterized by various spectroscopic techniques (IR, electronic, NMR, and electron paramagnetic resonance (EPR)) and thermal as well as field-emission scanning electron micrographs (FE-SEM) studies, and the molecular structures of 3 and 4 were determined by single crystal X-ray diffraction. The EPR spectrum of the polymer supported V(IV)O-complex 1 is characteristic of magnetically diluted V(IV)O-complexes, the resolved EPR pattern indicating that the V(IV)O-centers are well dispersed in the polymer matrix. A good (51)V NMR spectrum could also be measured with 4 suspended in dimethyl sulfoxide (DMSO), the chemical shift (-503 ppm) being compatible with a VO(2)(+)-center and a N,O binding set. The catalytic oxidative desulfurization of organosulfur compounds thiophene, dibenzothiophene, benzothiophene, and 2-methyl thiophene (model of fuel diesel) was carried out using complexes 1 and 2. The sulfur in model organosulfur compounds oxidizes to the corresponding sulfone in the presence of H(2)O(2). The systems 1 and 2 do not loose efficiency for sulfoxidation at least up to the third cycle of reaction, this indicating that they preserve their integrity under the conditions used. Plausible intermediates involved in these catalytic processes are established by UV-vis, EPR, (51)V NMR, and density functional theory (DFT) studies, and an outline of the mechanism is proposed. The (51)V NMR spectra

  16. Information-Processing Models and Curriculum Design

    Science.gov (United States)

    Calfee, Robert C.

    1970-01-01

    "This paper consists of three sections--(a) the relation of theoretical analyses of learning to curriculum design, (b) the role of information-processing models in analyses of learning processes, and (c) selected examples of the application of information-processing models to curriculum design problems." (Author)

  17. An accurate mobility model for the I-V characteristics of n-channel enhancement-mode MOSFETs with single-channel boron implantation

    International Nuclear Information System (INIS)

    Chingyuan Wu; Yeongwen Daih

    1985-01-01

    In this paper an analytical mobility model is developed for the I-V characteristics of n-channel enhancement-mode MOSFETs, in which the effects of the two-dimensional electric fields in the surface inversion channel and the parasitic resistances due to contact and interconnection are included. Most importantly, the developed mobility model easily takes the device structure and process into consideration. In order to demonstrate the capabilities of the developed model, the structure- and process-oriented parameters in the present mobility model are calculated explicitly for an n-channel enhancement-mode MOSFET with single-channel boron implantation. Moreover, n-channel MOSFETs with different channel lengths fabricated in a production line by using a set of test keys have been characterized and the measured mobilities have been compared to the model. Excellent agreement has been obtained for all ranges of the fabricated channel lengths, which strongly support the accuracy of the model. (author)

  18. Locating the Seventh Cervical Spinous Process: Development and Validation of a Multivariate Model Using Palpation and Personal Information.

    Science.gov (United States)

    Ferreira, Ana Paula A; Póvoa, Luciana C; Zanier, José F C; Ferreira, Arthur S

    2017-02-01

    The aim of this study was to develop and validate a multivariate prediction model, guided by palpation and personal information, for locating the seventh cervical spinous process (C7SP). A single-blinded, cross-sectional study at a primary to tertiary health care center was conducted for model development and temporal validation. One-hundred sixty participants were prospectively included for model development (n = 80) and time-split validation stages (n = 80). The C7SP was located using the thorax-rib static method (TRSM). Participants underwent chest radiography for assessment of the inner body structure located with TRSM and using radio-opaque markers placed over the skin. Age, sex, height, body mass, body mass index, and vertex-marker distance (D V-M ) were used to predict the distance from the C7SP to the vertex (D V-C7 ). Multivariate linear regression modeling, limits of agreement plot, histogram of residues, receiver operating characteristic curves, and confusion tables were analyzed. The multivariate linear prediction model for D V-C7 (in centimeters) was D V-C7 = 0.986D V-M + 0.018(mass) + 0.014(age) - 1.008. Receiver operating characteristic curves had better discrimination of D V-C7 (area under the curve = 0.661; 95% confidence interval = 0.541-0.782; P = .015) than D V-M (area under the curve = 0.480; 95% confidence interval = 0.345-0.614; P = .761), with respective cutoff points at 23.40 cm (sensitivity = 41%, specificity = 63%) and 24.75 cm (sensitivity = 69%, specificity = 52%). The C7SP was correctly located more often when using predicted D V-C7 in the validation sample than when using the TRSM in the development sample: n = 53 (66%) vs n = 32 (40%), P information. Copyright © 2016. Published by Elsevier Inc.

  19. Computer simulation for precipitation process of θ (Ni3V) and γ' (Ni3Al) with microscopic phase-field method

    Institute of Scientific and Technical Information of China (English)

    HOU Hua; XU Hong; ZHAO Yu-hong

    2006-01-01

    With the microscopic phase-field model, the precipitation process of aged alloys was explored by computer simulation, which could clarify some discussional views on the precipitation mechanisms of alloys. The precipitation process of Ni75Al2.5V22.5 alloy was studied. From the simulated atomic pictures, calculated order parameters and volume fraction of different precipitates, it's found that the θ ordered phase precipitates earlier than γ' ordered phase by congruent ordering+spinodal decomposition mechanism and thus produces a nonstoicheometric θ single ordered phase. Then, the nonstoicheometric γ' phase is precipitated by a non-classical nucleation and growth mechanism at the APBS of θ phases. Meanwhile, both of them transform to stoicheometric ordered phases.

  20. The study of production processes of nucleons, γ-quanta and neutral strange particles in pp- and pn-interactions at 300 GeV

    International Nuclear Information System (INIS)

    Turumov, Eh.

    2001-03-01

    Processes of nucleons, γ-quanta and neutral strange particles in pp- and pn-interactions at 300 GeV in 4π-geometry were studied by the use of neon-hydrogen bubble chamber. The average multiplicity of γ-quanta, fast neutrons, K 0 s -mesons and λ 0 ( λ 0 bar)-hyperons were defined and their energy spectra were analyzed. Experimental results were compared with Lund and dual parton model predictions. (author)

  1. The Procedural Queer: Substantive Due Process, "Lawrence v. Texas," and Queer Rhetorical Futures

    Science.gov (United States)

    Campbell, Peter Odell

    2012-01-01

    This essay discusses Justice Anthony M. Kennedy's choice to foreground arguments from due process rather than equal protection in the majority opinion in Lawrence v. Texas. Kennedy's choice can realize constitutional legal doctrine that is more consistent with radical queer politics than arguments from equal protection. Unlike some recent…

  2. Measuring similarity between business process models

    NARCIS (Netherlands)

    Dongen, van B.F.; Dijkman, R.M.; Mendling, J.

    2007-01-01

    Quality aspects become increasingly important when business process modeling is used in a large-scale enterprise setting. In order to facilitate a storage without redundancy and an efficient retrieval of relevant process models in model databases it is required to develop a theoretical understanding

  3. Model medication management process in Australian nursing homes using business process modeling.

    Science.gov (United States)

    Qian, Siyu; Yu, Ping

    2013-01-01

    One of the reasons for end user avoidance or rejection to use health information systems is poor alignment of the system with healthcare workflow, likely causing by system designers' lack of thorough understanding about healthcare process. Therefore, understanding the healthcare workflow is the essential first step for the design of optimal technologies that will enable care staff to complete the intended tasks faster and better. The often use of multiple or "high risk" medicines by older people in nursing homes has the potential to increase medication error rate. To facilitate the design of information systems with most potential to improve patient safety, this study aims to understand medication management process in nursing homes using business process modeling method. The paper presents study design and preliminary findings from interviewing two registered nurses, who were team leaders in two nursing homes. Although there were subtle differences in medication management between the two homes, major medication management activities were similar. Further field observation will be conducted. Based on the data collected from observations, an as-is process model for medication management will be developed.

  4. 78 FR 71032 - Foreign Tire Sales, Inc., Denial of Petition for Decision of Inconsequential Noncompliance

    Science.gov (United States)

    2013-11-27

    ...-0063; Notice 2] Foreign Tire Sales, Inc., Denial of Petition for Decision of Inconsequential...). ACTION: Denial of Petition. SUMMARY: Foreign Tire Sales, Inc. (FTS), as importer for ProMeter brand medium truck radial replacement tires manufactured by Shandlong Linglong Rubber Company Limited, has...

  5. Simulation Modeling of Software Development Processes

    Science.gov (United States)

    Calavaro, G. F.; Basili, V. R.; Iazeolla, G.

    1996-01-01

    A simulation modeling approach is proposed for the prediction of software process productivity indices, such as cost and time-to-market, and the sensitivity analysis of such indices to changes in the organization parameters and user requirements. The approach uses a timed Petri Net and Object Oriented top-down model specification. Results demonstrate the model representativeness, and its usefulness in verifying process conformance to expectations, and in performing continuous process improvement and optimization.

  6. Social software for business process modeling

    NARCIS (Netherlands)

    Koschmider, A.; Song, M.S.; Reijers, H.A.

    2010-01-01

    Formal models of business processes are used for a variety of purposes. But where the elicitation of the characteristics of a business process usually takes place in a collaborative fashion, the building of the final, formal process model is done mostly by a single person. This article presents the

  7. Validation of ASTEC v2.0 corium jet fragmentation model using FARO experiments

    International Nuclear Information System (INIS)

    Hermsmeyer, S.; Pla, P.; Sangiorgi, M.

    2015-01-01

    Highlights: • Model validation base extended to six FARO experiments. • Focus on the calculation of the fragmented particle diameter. • Capability and limits of the ASTEC fragmentation model. • Sensitivity analysis of model outputs. - Abstract: ASTEC is an integral code for the prediction of Severe Accidents in Nuclear Power Plants. As such, it needs to cover all physical processes that could occur during accident progression, yet keeping its models simple enough for the ensemble to stay manageable and produce results within an acceptable time. The present paper is concerned with the validation of the Corium jet fragmentation model of ASTEC v2.0 rev3 by means of a selection of six experiments carried out within the FARO facility. The different conditions applied within these six experiments help to analyse the model behaviour in different situations and to expose model limits. In addition to comparing model outputs with experimental measurements, sensitivity analyses are applied to investigate the model. Results of the paper are (i) validation runs, accompanied by an identification of situations where the implemented fragmentation model does not match the experiments well, and discussion of results; (ii) its special attention to the models calculating the diameter of fragmented particles, the identification of a fault in one model implemented, and the discussion of simplification and ad hoc modification to improve the model fit; and, (iii) an investigation of the sensitivity of predictions towards inputs and parameters. In this way, the paper offers a thorough investigation of the merit and limitation of the fragmentation model used in ASTEC

  8. Modeling business processes: theoretical and practical aspects

    Directory of Open Access Journals (Sweden)

    V.V. Dubininа

    2015-06-01

    Full Text Available The essence of process-oriented enterprise management has been examined in the article. The content and types of information technology have been analyzed in the article, due to the complexity and differentiation of existing methods, as well as the specificity of language, terminology of the enterprise business processes modeling. The theoretical aspects of business processes modeling have been reviewed and the modern traditional modeling techniques received practical application in the visualization model of retailers activity have been studied in the article. In the process of theoretical analysis of the modeling methods found that UFO-toolkit method that has been developed by Ukrainian scientists due to it systemology integrated opportunities, is the most suitable for structural and object analysis of retailers business processes. It was designed visualized simulation model of the business process "sales" as is" of retailers using a combination UFO-elements with the aim of the further practical formalization and optimization of a given business process.

  9. Transport and mass exchange processes in sand and gravel aquifers (v.1)

    International Nuclear Information System (INIS)

    Moltyaner, G.

    1990-01-01

    The objectives of this conference were to exchange information on promising field measurement techniques used for the characterization of spatial variability of geologic formations and on new methods used for quantifying the effect of spatial variability on groundwater flow and transport of materials; to discuss novel developments in the theory of transport processes and simulation methods; and to present views and opinions on future initiatives and directions in the design of large-scale field tracer experiments and the development of conceptual and mathematical models of transport and mass exchange processes. The 46 papers presented in these proceedings are divided into six sections: field studies of transport processes; groundwater tracers and novel field measurement techniques; promising methods and field measurement techniques for quantifying the effect of geological heterogeneities on groundwater flow and transport; novel developments in the theory of transport processes; numerical modelling of transport and mass exchange processes; and field and modelling studies of mass exchange processes. (L.L.)

  10. Coupling a three-dimensional subsurface flow and transport model with a land surface model to simulate stream–aquifer–land interactions (CP v1.0

    Directory of Open Access Journals (Sweden)

    G. Bisht

    2017-12-01

    Full Text Available A fully coupled three-dimensional surface and subsurface land model is developed and applied to a site along the Columbia River to simulate three-way interactions among river water, groundwater, and land surface processes. The model features the coupling of the Community Land Model version 4.5 (CLM4.5 and a massively parallel multiphysics reactive transport model (PFLOTRAN. The coupled model, named CP v1.0, is applied to a 400 m × 400 m study domain instrumented with groundwater monitoring wells along the Columbia River shoreline. CP v1.0 simulations are performed at three spatial resolutions (i.e., 2, 10, and 20 m over a 5-year period to evaluate the impact of hydroclimatic conditions and spatial resolution on simulated variables. Results show that the coupled model is capable of simulating groundwater–river-water interactions driven by river stage variability along managed river reaches, which are of global significance as a result of over 30 000 dams constructed worldwide during the past half-century. Our numerical experiments suggest that the land-surface energy partitioning is strongly modulated by groundwater–river-water interactions through expanding the periodically inundated fraction of the riparian zone, and enhancing moisture availability in the vadose zone via capillary rise in response to the river stage change. Meanwhile, CLM4.5 fails to capture the key hydrologic process (i.e., groundwater–river-water exchange at the site, and consequently simulates drastically different water and energy budgets. Furthermore, spatial resolution is found to significantly impact the accuracy of estimated the mass exchange rates at the boundaries of the aquifer, and it becomes critical when surface and subsurface become more tightly coupled with groundwater table within 6 to 7 meters below the surface. Inclusion of lateral subsurface flow influenced both the surface energy budget and subsurface transport processes as a result

  11. Coupling a three-dimensional subsurface flow and transport model with a land surface model to simulate stream-aquifer-land interactions (CP v1.0)

    Science.gov (United States)

    Bisht, Gautam; Huang, Maoyi; Zhou, Tian; Chen, Xingyuan; Dai, Heng; Hammond, Glenn E.; Riley, William J.; Downs, Janelle L.; Liu, Ying; Zachara, John M.

    2017-12-01

    A fully coupled three-dimensional surface and subsurface land model is developed and applied to a site along the Columbia River to simulate three-way interactions among river water, groundwater, and land surface processes. The model features the coupling of the Community Land Model version 4.5 (CLM4.5) and a massively parallel multiphysics reactive transport model (PFLOTRAN). The coupled model, named CP v1.0, is applied to a 400 m × 400 m study domain instrumented with groundwater monitoring wells along the Columbia River shoreline. CP v1.0 simulations are performed at three spatial resolutions (i.e., 2, 10, and 20 m) over a 5-year period to evaluate the impact of hydroclimatic conditions and spatial resolution on simulated variables. Results show that the coupled model is capable of simulating groundwater-river-water interactions driven by river stage variability along managed river reaches, which are of global significance as a result of over 30 000 dams constructed worldwide during the past half-century. Our numerical experiments suggest that the land-surface energy partitioning is strongly modulated by groundwater-river-water interactions through expanding the periodically inundated fraction of the riparian zone, and enhancing moisture availability in the vadose zone via capillary rise in response to the river stage change. Meanwhile, CLM4.5 fails to capture the key hydrologic process (i.e., groundwater-river-water exchange) at the site, and consequently simulates drastically different water and energy budgets. Furthermore, spatial resolution is found to significantly impact the accuracy of estimated the mass exchange rates at the boundaries of the aquifer, and it becomes critical when surface and subsurface become more tightly coupled with groundwater table within 6 to 7 meters below the surface. Inclusion of lateral subsurface flow influenced both the surface energy budget and subsurface transport processes as a result of river-water intrusion into the

  12. Modeling closed nuclear fuel cycles processes

    Energy Technology Data Exchange (ETDEWEB)

    Shmidt, O.V. [A.A. Bochvar All-Russian Scientific Research Institute for Inorganic Materials, Rogova, 5a street, Moscow, 123098 (Russian Federation); Makeeva, I.R. [Zababakhin All-Russian Scientific Research Institute of Technical Physics, Vasiliev street 13, Snezhinsk, Chelyabinsk region, 456770 (Russian Federation); Liventsov, S.N. [Tomsk Polytechnic University, Tomsk, Lenin Avenue, 30, 634050 (Russian Federation)

    2016-07-01

    Computer models of processes are necessary for determination of optimal operating conditions for closed nuclear fuel cycle (NFC) processes. Computer models can be quickly changed in accordance with new and fresh data from experimental research. 3 kinds of process simulation are necessary. First, the VIZART software package is a balance model development used for calculating the material flow in technological processes. VIZART involves taking into account of equipment capacity, transport lines and storage volumes. Secondly, it is necessary to simulate the physico-chemical processes that are involved in the closure of NFC. The third kind of simulation is the development of software that allows the optimization, diagnostics and control of the processes which implies real-time simulation of product flows on the whole plant or on separate lines of the plant. (A.C.)

  13. PISCES-v2: an ocean biogeochemical model for carbon and ecosystem studies

    Directory of Open Access Journals (Sweden)

    O. Aumont

    2015-08-01

    of marine ecosystems (phytoplankton, microzooplankton and mesozooplankton and the biogeochemical cycles of carbon and of the main nutrients (P, N, Fe, and Si. The model is intended to be used for both regional and global configurations at high or low spatial resolutions as well as for short-term (seasonal, interannual and long-term (climate change, paleoceanography analyses. There are 24 prognostic variables (tracers including two phytoplankton compartments (diatoms and nanophytoplankton, two zooplankton size classes (microzooplankton and mesozooplankton and a description of the carbonate chemistry. Formulations in PISCES-v2 are based on a mixed Monod–quota formalism. On the one hand, stoichiometry of C / N / P is fixed and growth rate of phytoplankton is limited by the external availability in N, P and Si. On the other hand, the iron and silicon quotas are variable and the growth rate of phytoplankton is limited by the internal availability in Fe. Various parameterizations can be activated in PISCES-v2, setting, for instance, the complexity of iron chemistry or the description of particulate organic materials. So far, PISCES-v2 has been coupled to the Nucleus for European Modelling of the Ocean (NEMO and Regional Ocean Modeling System (ROMS systems. A full description of PISCES-v2 and of its optional functionalities is provided here. The results of a quasi-steady-state simulation are presented and evaluated against diverse observational and satellite-derived data. Finally, some of the new functionalities of PISCES-v2 are tested in a series of sensitivity experiments.

  14. Eye Tracking Meets the Process of Process Modeling: a Visual Analytic Approach

    DEFF Research Database (Denmark)

    Burattin, Andrea; Kaiser, M.; Neurauter, Manuel

    2017-01-01

    Research on the process of process modeling (PPM) studies how process models are created. It typically uses the logs of the interactions with the modeling tool to assess the modeler’s behavior. In this paper we suggest to introduce an additional stream of data (i.e., eye tracking) to improve the ...

  15. Modeling biochemical transformation processes and information processing with Narrator.

    Science.gov (United States)

    Mandel, Johannes J; Fuss, Hendrik; Palfreyman, Niall M; Dubitzky, Werner

    2007-03-27

    Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs), which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Narrator is a flexible and intuitive systems biology tool. It is

  16. Selective Laser Melting Produced Ti-6Al-4V: Post-Process Heat Treatments to Achieve Superior Tensile Properties

    Directory of Open Access Journals (Sweden)

    Gerrit M. Ter Haar

    2018-01-01

    Full Text Available Current post-process heat treatments applied to selective laser melting produced Ti-6Al-4V do not achieve the same microstructure and therefore superior tensile behaviour of thermomechanical processed wrought Ti-6Al-4V. Due to the growing demand for selective laser melting produced parts in industry, research and development towards improved mechanical properties is ongoing. This study is aimed at developing post-process annealing strategies to improve tensile behaviour of selective laser melting produced Ti-6Al-4V parts. Optical and electron microscopy was used to study α grain morphology as a function of annealing temperature, hold time and cooling rate. Quasi-static uniaxial tensile tests were used to measure tensile behaviour of different annealed parts. It was found that elongated α’/α grains can be fragmented into equiaxial grains through applying a high temperature annealing strategy. It is shown that bi-modal microstructures achieve a superior tensile ductility to current heat treated selective laser melting produced Ti-6Al-4V samples.

  17. The Formalization of the Business Process Modeling Goals

    Directory of Open Access Journals (Sweden)

    Ligita Bušinska

    2016-10-01

    Full Text Available In business process modeling the de facto standard BPMN has emerged. However, the applications of this notation have many subsets of elements and various extensions. Also, BPMN still coincides with many other modeling languages, forming a large set of available options for business process modeling languages and dialects. While, in general, the goal of modelers is a central notion in the choice of modeling languages and notations, in most researches that propose guidelines, techniques, and methods for business process modeling language evaluation and/or selection, the business process modeling goal is not formalized and not transparently taken into account. To overcome this gap, and to explicate and help to handle business process modeling complexity, the approach to formalize the business process modeling goal, and the supporting three dimensional business process modeling framework, are proposed.

  18. Fuel fragmentation model advances using TEXAS-V

    Energy Technology Data Exchange (ETDEWEB)

    Corradini, M.L.; El-Beshbeeshy, M.; Nilsuwankowsit, S.; Tang, J. [Wisconsin Univ., Madison, WI (United States). Dept. of Nuclear Engineering and Engineering Physics

    1998-01-01

    Because an energetic fuel-coolant interaction may be a safety hazard, experiments are being conducted to investigate the fuel-coolant mixing/quenching process (FARO) as well as the energetics of vapor explosion propagation for high temperature fuel melt simulants (KROTOS, WFCI, ZrEX). In both types of experiments, the dynamic breakup of the fuel is one of the key aspects that must be fundamentally understood to better estimate the magnitude of the mixing/quenching process or the explosion energetics. To aid our understanding the TEXAS fuel-coolant interaction computer model has been developed and is being used to analyze these experiments. Recently, the models for dynamic fuel fragmentation during the mixing and explosion phases of the FCI have been improved by further insights into these processes. The purpose of this paper is to describe these enhancements and to demonstrate their improvements by analysis of particular JRC FCI data. (author)

  19. Pre-Processing and Modeling Tools for Bigdata

    Directory of Open Access Journals (Sweden)

    Hashem Hadi

    2016-09-01

    Full Text Available Modeling tools and operators help the user / developer to identify the processing field on the top of the sequence and to send into the computing module only the data related to the requested result. The remaining data is not relevant and it will slow down the processing. The biggest challenge nowadays is to get high quality processing results with a reduced computing time and costs. To do so, we must review the processing sequence, by adding several modeling tools. The existing processing models do not take in consideration this aspect and focus on getting high calculation performances which will increase the computing time and costs. In this paper we provide a study of the main modeling tools for BigData and a new model based on pre-processing.

  20. Kopernik : modeling business processes for digital customers

    OpenAIRE

    Estañol Lamarca, Montserrat; Castro, Manuel; Díaz-Montenegro, Sylvia; Teniente López, Ernest

    2016-01-01

    This paper presents the Kopernik methodology for modeling business processes for digital customers. These processes require a high degree of flexibility in the execution of their tasks or actions. We achieve this by using the artifact-centric approach to process modeling and the use of condition-action rules. The processes modeled following Kopernik can then be implemented in an existing commercial tool, Balandra.

  1. The probabilistic model of the process mixing of animal feed ingredients into a continuous mixer-reactor

    Directory of Open Access Journals (Sweden)

    L. I. Lytkina

    2016-01-01

    Full Text Available A mathematical model of the polydisperse medium mixing process reflects its stochastic features in the form of uneven distribution of phase elements on the time of their presence in apparatus, particle size, ripple retention of the apparatus, random distribution of the material and thermal phase flows of the working volume, heterogeneity of the medium physical- and chemical properties, complicated by chemical reaction. For the mathematical description of the mixing process of animal feed ingredients in the presence of chemical reaction the system of differential equations of Academician V.V. Kafarov was used. Proposed by him hypothesis based on the theory of Markov’s processes stating that "any multicomponent mixture can be considered as the result of an iterative process of mixing the two components to achieve the desired uniformity of all the ingredients in the mixture" allows us to consider a process of mixing binary composition in a paddle mixer in the form of differential equations of two ingredients concentration numerous changes until it becomes a homogenous mixture. It was found out that the mixing process of the two-component mixture is determined in a paddle mixer with a constant mixing speed and a limit (equilibrium dispersion of the ingredients in the mixture i.e. with its uniformity. Adjustment of the model parameters was carried out according to the results of experimental studies on mixing the crushed wheat with metallomagnetic impurity, which was a key (indicator component. According to the best values of the constant of the continuous mixing speed and the equilibrium disperse values of the ingredients contents, the mathematical model parameters identification was carried out. The results obtained are used to develop a new generation mixer design.

  2. Specification of e-business process model for PayPal online payment process using Reo

    OpenAIRE

    Xie, M.

    2005-01-01

    textabstractE-business process modeling allows business analysts to better understand and analyze the business processes, and eventually to use software systems to automate (parts of) these business processes to achieve higher profit. To support e-business process modeling, many business process modeling languages have been used as tools. However, many existing business process modeling languages lack (a) formal semantics, (b) formal computational model, and (c) an integrated view of the busi...

  3. Process mining using BPMN: relating event logs and process models

    NARCIS (Netherlands)

    Kalenkova, A.A.; van der Aalst, W.M.P.; Lomazova, I.A.; Rubin, V.A.

    2017-01-01

    Process-aware information systems (PAIS) are systems relying on processes, which involve human and software resources to achieve concrete goals. There is a need to develop approaches for modeling, analysis, improvement and monitoring processes within PAIS. These approaches include process mining

  4. Process mining using BPMN : relating event logs and process models

    NARCIS (Netherlands)

    Kalenkova, A.A.; Aalst, van der W.M.P.; Lomazova, I.A.; Rubin, V.A.

    2015-01-01

    Process-aware information systems (PAIS) are systems relying on processes, which involve human and software resources to achieve concrete goals. There is a need to develop approaches for modeling, analysis, improvement and monitoring processes within PAIS. These approaches include process mining

  5. A Comparative of business process modelling techniques

    Science.gov (United States)

    Tangkawarow, I. R. H. T.; Waworuntu, J.

    2016-04-01

    In this era, there is a lot of business process modeling techniques. This article is the research about differences of business process modeling techniques. For each technique will explain about the definition and the structure. This paper presents a comparative analysis of some popular business process modelling techniques. The comparative framework is based on 2 criteria: notation and how it works when implemented in Somerleyton Animal Park. Each technique will end with the advantages and disadvantages. The final conclusion will give recommend of business process modeling techniques that easy to use and serve the basis for evaluating further modelling techniques.

  6. Piecewise deterministic processes in biological models

    CERN Document Server

    Rudnicki, Ryszard

    2017-01-01

    This book presents a concise introduction to piecewise deterministic Markov processes (PDMPs), with particular emphasis on their applications to biological models. Further, it presents examples of biological phenomena, such as gene activity and population growth, where different types of PDMPs appear: continuous time Markov chains, deterministic processes with jumps, processes with switching dynamics, and point processes. Subsequent chapters present the necessary tools from the theory of stochastic processes and semigroups of linear operators, as well as theoretical results concerning the long-time behaviour of stochastic semigroups induced by PDMPs and their applications to biological models. As such, the book offers a valuable resource for mathematicians and biologists alike. The first group will find new biological models that lead to interesting and often new mathematical questions, while the second can observe how to include seemingly disparate biological processes into a unified mathematical theory, and...

  7. Computer Aided Continuous Time Stochastic Process Modelling

    DEFF Research Database (Denmark)

    Kristensen, N.R.; Madsen, Henrik; Jørgensen, Sten Bay

    2001-01-01

    A grey-box approach to process modelling that combines deterministic and stochastic modelling is advocated for identification of models for model-based control of batch and semi-batch processes. A computer-aided tool designed for supporting decision-making within the corresponding modelling cycle...

  8. Phosphates (V) recovery from phosphorus mineral fertilizers industry wastewater by continuous struvite reaction crystallization process.

    Science.gov (United States)

    Hutnik, Nina; Kozik, Anna; Mazienczuk, Agata; Piotrowski, Krzysztof; Wierzbowska, Boguslawa; Matynia, Andrzej

    2013-07-01

    Continuous DT MSMPR (Draft Tube Mixed Suspension Mixed Product Removal) crystallizer was provided with typical wastewater from phosphorus mineral fertilizers industry (pH < 4, 0.445 mass % of PO4(3-), inorganic impurities presence), dissolved substrates (magnesium and ammonium chlorides) and solution alkalising the environment of struvite MgNH4PO4·6H2O reaction crystallization process. Research ran in constant temperature 298 K assuming stoichiometric proportions of substrates or 20% excess of magnesium ions. Influence of pH (8.5-10) and mean residence time (900-3600 s) on product size distribution, its chemical composition, crystals shape, size-homogeneity and process kinetics was identified. Crystals of mean size ca. 25-37 μm and homogeneity CV 70-83% were produced. The largest crystals, of acceptable homogeneity, were produced using 20% excess of magnesium ions, pH 9 and mean residence time 3600 s. Under these conditions nucleation rate did not exceed 9 × 10(7) 1/(s m(3)) according to SIG (Size Independent Growth) MSMPR kinetic model. Linear crystal growth rate was 4.27 × 10(-9) m/s. Excess of magnesium ions influenced struvite reaction crystallization process yield advantageously. Concentration of phosphate(V) ions decreased from 0.445 to 9.2 × 10(-4) mass %. This can be regarded as a very good process result. In product crystals, besides main component - struvite, all impurities from wastewater were detected analytically. Copyright © 2013 Elsevier Ltd. All rights reserved.

  9. Electron Processing at 50 eV of Terphenylthiol Self-Assembled Monolayers: Contributions of Primary and Secondary Electrons.

    Science.gov (United States)

    Houplin, Justine; Dablemont, Céline; Sala, Leo; Lafosse, Anne; Amiaud, Lionel

    2015-12-22

    Aromatic self-assembled monolayers (SAMs) can serve as platforms for development of supramolecular assemblies driven by surface templates. For many applications, electron processing is used to locally reinforce the layer. To achieve better control of the irradiation step, chemical transformations induced by electron impact at 50 eV of terphenylthiol SAMs are studied, with these SAMs serving as model aromatic SAMs. High-resolution electron energy loss spectroscopy (HREELS) and electron-stimulated desorption (ESD) of neutral fragment measurements are combined to investigate electron-induced chemical transformation of the layer. The decrease of the CH stretching HREELS signature is mainly attributed to dehydrogenation, without a noticeable hybridization change of the hydrogenated carbon centers. Its evolution as a function of the irradiation dose gives an estimate of the effective hydrogen content loss cross-section, σ = 2.7-4.7 × 10(-17) cm(2). Electron impact ionization is the major primary mechanism involved, with the impact electronic excitation contributing only marginally. Therefore, special attention is given to the contribution of the low-energy secondary electrons to the induced chemistry. The effective cross-section related to dissociative secondary electron attachment at 6 eV is estimated to be 1 order of magnitude smaller. The 1 eV electrons do not induce significant chemical modification for a 2.5 mC cm(-2) dose, excluding their contribution.

  10. 75 FR 32536 - Foreign Tire Sales, Inc., Receipt of Petition for Decision of Inconsequential Noncompliance

    Science.gov (United States)

    2010-06-08

    ...-0063; Notice 1] Foreign Tire Sales, Inc., Receipt of Petition for Decision of Inconsequential Noncompliance Foreign Tire Sales, Inc. (FTS) \\1\\, as importer of record for ProMeter brand medium truck radial replacement tires manufactured by Shandlong Linglong Rubber Company Limited has determined that certain...

  11. Neuroscientific Model of Motivational Process

    OpenAIRE

    Kim, Sung-il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Rewa...

  12. Central vasopressin V1a receptors modulate neural processing in mothers facing intruder threat to pups

    OpenAIRE

    Caffrey, Martha K.; Nephew, Benjamin C.; Febo, Marcelo

    2009-01-01

    Vasopressin V1a receptors in the rat brain have been studied for their role in modulating aggression and anxiety. In the current study blood-oxygen-level-dependent (BOLD) functional MRI was used to test whether V1a receptors modulate neural processing in the maternal brain when dams are exposed to a male intruder. Primiparous females were given an intracerebroventricular (ICV) injection of vehicle or V1a receptor antagonist ([deamino-Pen1, O-Me-Try, Arg8]-Vasopressin, 125 ng/10 μL) 90-120 min...

  13. Measures of Quality in Business Process Modelling

    Directory of Open Access Journals (Sweden)

    Radek Hronza

    2015-06-01

    Full Text Available Business process modelling and analysing is undoubtedly one of the most important parts of Applied (Business Informatics. Quality of business process models (diagrams is crucial for any purpose in this area. The goal of a process analyst’s work is to create generally understandable, explicit and error free models. If a process is properly described, created models can be used as an input into deep analysis and optimization. It can be assumed that properly designed business process models (similarly as in the case of correctly written algorithms contain characteristics that can be mathematically described. Besides it will be possible to create a tool that will help process analysts to design proper models. As part of this review will be conducted systematic literature review in order to find and analyse business process model’s design and business process model’s quality measures. It was found that mentioned area had already been the subject of research investigation in the past. Thirty-three suitable scietific publications and twenty-two quality measures were found. Analysed scientific publications and existing quality measures do not reflect all important attributes of business process model’s clarity, simplicity and completeness. Therefore it would be appropriate to add new measures of quality.

  14. Characteristics of the process K+p→K+anti ppp at 12 GeV/c

    International Nuclear Information System (INIS)

    Armstrong, T.A.; Frame, D.; Hughes, I.S.; Kumar, B.R.; Lewis, G.M.; Macallister, J.B.; Stewart, D.T.; Turnbull, R.M.

    1979-01-01

    Data are presented on the reaction K + p→K + anti ppp at 12 GeV/c from an experiment using the OMEGA spectrometer at CERN. A clear Λ(1520) signal is observed in the process K + p→anti Λ(1520)pp and angular distributions and correlations are presented for this process. The angular distributions for the reaction in which anti Λ(1520) is not produced show an appreciable backward K + peak. (Auth.)

  15. Study of dimuon pair production in e+ e- collisions from 196 - 202 GeV

    CERN Document Server

    Flacher, Henning

    2000-01-01

    In this thesis the electroweak process $e^{+} e^{-}$ --> $μ^{+} μ^{-}$ was studied and an inclusive and exclusive cross section were measured. Furthermore the forward-backward asymmetry Afb was determined from the exclusive event sample. The investigated data was recorded·with the detector ALEPH at centre-of-mass energies of 196, 200 and 202 GeV resulting in a total integrated luminosity of 208.1 pb-1. All the measured results are in good agreement with the Standard Model. From the measurements of total cross sections and angular distributions for all the two fermion processes at energies from 130 - 202 GeV limits on processes beyond the Standard Model were derived. For Contact Interactions they were found to be of the order of 10 TeV while for TeV-Scale Quantum Gravity a limit for the ultra-violet cut-off parameter of A::1 TeV could be derived.

  16. Time dependent approach of TeV blazars based on a model of inhomogeneous stratified jet

    International Nuclear Information System (INIS)

    Boutelier, T.

    2009-05-01

    The study of the emission and variability mechanisms of TeV blazars has been the subject of intensive research for years. The homogeneous one-zone model commonly used is puzzling since it yields very high Lorentz factor, in contradiction with other observational evidences. In this work, I describe a new time dependent multi-zone approach, in the framework of the two-flow model. I compute the emission of a full jet, where relativistic electron-positron pairs distributed in pileup propagate. The evolution and the emission of the plasma is computed taking into account a turbulent heating term, some radiative cooling, and a pair production term due to photo-annihilation process. Applied to PKS 2155-304, the model allows the reproduction of the full spectra, as well as the simultaneous multi wavelength variability, with a relatively small Lorentz factor. The variability is explained by the instability of the pair creation process. Nonetheless, the value is still high to agree with other observational evidences in radio. Hence, I show in the last part of this work how to conciliate high Lorentz factor with the absence of apparent superluminal movement in radio, by taking into account the effect of the opening angle on the appearance of relativistic jets. (author)

  17. Disability intervention model for older adults with arthritis: an integration of theory of symptom management and disablement process model.

    Science.gov (United States)

    Shin, So Young

    2014-12-01

    To evolve a management plan for rheumatoid arthritis, it is necessary to understand the patient's symptom experience and disablement process. This paper aims to introduce and critique two models as a conceptual foundation from which to construct a new model for arthritis care. A Disability Intervention Model for Older Adults with Arthritis includes three interrelated concepts of symptom experience, symptom management strategies, and symptom outcomes that correspond to the Theory of Symptom Management. These main concepts influence or are influenced by contextual factors that are situated within the domains of person, environment, and health/illness. It accepts the bidirectional, complex, dynamic interactions among all components within the model representing the comprehensive aspects of the disablement process and its interventions in older adults with rheumatoid arthritis. In spite of some limitations such as confusion or complexity within the model, the Disability Intervention Model for Older Adults with Arthritis has strengths in that it encompasses the majority of the concepts of the two models, attempts to compensate for the limitations of the two models, and aims to understand the impact of rheumatoid arthritis on a patient's physical, cognitive, and emotional health status, socioeconomic status, and well-being. Therefore, it can be utilized as a guiding theoretical framework for arthritis care and research to improve the functional status of older adults with rheumatoid arthritis. Copyright © 2014. Published by Elsevier B.V.

  18. Collisional-radiative model including recombination processes for W27+ ion★

    Science.gov (United States)

    Murakami, Izumi; Sasaki, Akira; Kato, Daiji; Koike, Fumihiro

    2017-10-01

    We have constructed a collisional-radiative (CR) model for W27+ ions including 226 configurations with n ≤ 9 and ł ≤ 5 for spectroscopic diagnostics. We newly include recombination processes in the model and this is the first result of extreme ultraviolet spectrum calculated for recombining plasma component. Calculated spectra in 40-70 Å range in ionizing and recombining plasma components show similar 3 strong lines and 1 line weak in recombining plasma component at 45-50 Å and many weak lines at 50-65 Å for both components. Recombination processes do not contribute much to the spectrum at around 60 Å for W27+ ion. Dielectronic satellite lines are also minor contribution to the spectrum of recombining plasma component. Dielectronic recombination (DR) rate coefficient from W28+ to W27+ ions is also calculated with the same atomic data in the CR model. We found that larger set of energy levels including many autoionizing states gave larger DR rate coefficients but our rate agree within factor 6 with other works at electron temperature around 1 keV in which W27+ and W28+ ions are usually observed in plasmas. Contribution to the Topical Issue "Atomic and Molecular Data and their Applications", edited by Gordon W.F. Drake, Jung-Sik Yoon, Daiji Kato, and Grzegorz Karwasz.

  19. Microstructure and tensile properties after thermohydrogen processing of Ti-6 Al-4V.

    Science.gov (United States)

    Guitar, A; Vigna, G; Luppo, M I

    2009-04-01

    Thermohydrogen processing (THP), a technique in which hydrogen is used as a temporary alloying element, can refine the microstructure and improve the final mechanical properties of the Ti-6 Al-4V alloy. THP allows microstructural modification of titanium alloys near net shape such as biomaterial components obtained by powder metallurgy and castings, since it does not require mechanical working. Two THP, called THP-A and THP-B, have been evaluated in samples of Ti-6Al-4V with a coarse and lamellar microstructure typical of castings and powder metallurgy. The THP-A is based in the eutectoid decomposition of the beta(H) phase to alpha phase and hydride phase. The THP-B is based in the isothermal decomposition of alpha('') martensite phase, obtained by quenching of hydrogenated samples. The refinement of the microstructure due to THP has been evaluated by means of optical and electron microscopy. Tensile tests showed that while both processes were able to increase the strength of the alloy as compared with the starting material, the ductility in samples subjected to THP-B was severely reduced.

  20. Modeling of ETL-Processes and Processed Information in Clinical Data Warehousing.

    Science.gov (United States)

    Tute, Erik; Steiner, Jochen

    2018-01-01

    Literature describes a big potential for reuse of clinical patient data. A clinical data warehouse (CDWH) is a means for that. To support management and maintenance of processes extracting, transforming and loading (ETL) data into CDWHs as well as to ease reuse of metadata between regular IT-management, CDWH and secondary data users by providing a modeling approach. Expert survey and literature review to find requirements and existing modeling techniques. An ETL-modeling-technique was developed extending existing modeling techniques. Evaluation by exemplarily modeling existing ETL-process and a second expert survey. Nine experts participated in the first survey. Literature review yielded 15 included publications. Six existing modeling techniques were identified. A modeling technique extending 3LGM2 and combining it with openEHR information models was developed and evaluated. Seven experts participated in the evaluation. The developed approach can help in management and maintenance of ETL-processes and could serve as interface between regular IT-management, CDWH and secondary data users.

  1. Simulation Models of Human Decision-Making Processes

    Directory of Open Access Journals (Sweden)

    Nina RIZUN

    2014-10-01

    Full Text Available The main purpose of the paper is presentation of the new concept of human decision-making process modeling via using the analogy with Automatic Control Theory. From the author's point of view this concept allows to develop and improve the theory of decision-making in terms of the study and classification of specificity of the human intellectual processes in different conditions. It was proved that the main distinguishing feature between the Heuristic / Intuitive and Rational Decision-Making Models is the presence of so-called phenomenon of "enrichment" of the input information with human propensity, hobbies, tendencies, expectations, axioms and judgments, presumptions or bias and their justification. In order to obtain additional knowledge about the basic intellectual processes as well as the possibility of modeling the decision results in various parameters characterizing the decision-maker, the complex of the simulation models was developed. These models are based on the assumptions that:  basic intellectual processes of the Rational Decision-Making Model can be adequately simulated and identified by the transient processes of the proportional-integral-derivative controller; basic intellectual processes of the Bounded Rationality and Intuitive Models can be adequately simulated and identified by the transient processes of the nonlinear elements.The taxonomy of the most typical automatic control theory elements and their compliance with certain decision-making models with a point of view of decision-making process specificity and decision-maker behavior during a certain time of professional activity was obtained.

  2. Search for the Standard Model Higgs Boson in ZH → v$\\bar{v}$b$\\bar{b}$ channel in p$\\bar{p}$ collisions at √s= 1.96 TeV

    Energy Technology Data Exchange (ETDEWEB)

    Dubey, Abhinav [Univ. of Delhi, New Delhi (India)

    2011-01-01

    A search for the standard model Higgs boson is performed in 5.2 fb-1 of p$\\bar{p}$ collisions at p √s = 1.96 TeV, collected with the DØ detector at the Fermilab Tevatron. The final state considered is a pair of b jets with large missing transverse energy, as expected from p$\\bar{p}$→ ZH → v$\\bar{v}$b$\\bar{b}$ production. The search is also sensitive to the WH → ℓvb$\\bar{b}$ channel, where the charged lepton is not identified. Boosted decision trees are used to discriminate signal from background. Good agreement is observed between data and expected backgrounds, and, for a Higgs-boson mass of 115 GeV, a limit is set at 95% C.L. on the cross section multiplied by branching fraction of (p$\\bar{p}$ → (Z/W)H) × (H → b$\\bar{b}$) that is a factor 4.57 expected and 3.73 observed larger than the value expected from the standard model.

  3. User-guided discovery of declarative process models

    NARCIS (Netherlands)

    Maggi, F.M.; Mooij, A.J.; Aalst, van der W.M.P.; Chawla, N.; King, I.; Sperduti, A.

    2011-01-01

    Process mining techniques can be used to effectively discover process models from logs with example behaviour. Cross-correlating a discovered model with information in the log can be used to improve the underlying process. However, existing process discovery techniques have two important drawbacks.

  4. Deferred Action: Theoretical model of process architecture design for emergent business processes

    Directory of Open Access Journals (Sweden)

    Patel, N.V.

    2007-01-01

    Full Text Available E-Business modelling and ebusiness systems development assumes fixed company resources, structures, and business processes. Empirical and theoretical evidence suggests that company resources and structures are emergent rather than fixed. Planning business activity in emergent contexts requires flexible ebusiness models based on better management theories and models . This paper builds and proposes a theoretical model of ebusiness systems capable of catering for emergent factors that affect business processes. Drawing on development of theories of the ‘action and design’class the Theory of Deferred Action is invoked as the base theory for the theoretical model. A theoretical model of flexible process architecture is presented by identifying its core components and their relationships, and then illustrated with exemplar flexible process architectures capable of responding to emergent factors. Managerial implications of the model are considered and the model’s generic applicability is discussed.

  5. INCOSE Systems Engineering Handbook v3.2: Improving the Process for SE Practitioners

    Energy Technology Data Exchange (ETDEWEB)

    R. Douglas Hamelin; David D. Walden; Michael E. Krueger

    2010-07-01

    The INCOSE Systems Engineering Handbook is the official INCOSE reference document for understanding systems engineering (SE) methods and conducting SE activities. Over the years, the Handbook has evolved to accommodate advances in the SE discipline and now serves as the basis for the Certified Systems Engineering Professional (CSEP) exam. Due to its evolution, the Handbook had become somewhat disjointed in its treatment and presentation of SE topics and was not aligned with the latest version of International Organization for Standardization (ISO)/International Electrotechnical Commission (IEC) 15288:2008, Systems and Software Engineering. As a result, numerous inconsistencies were identified that could confuse practitioners and directly impact the probability of success in passing the CSEP exam. Further, INCOSE leadership had previously submitted v3.1 of the Handbook to ISO/IEC for consideration as a Technical Report, but was told that the Handbook would have to be updated to conform with the terminology and structure of new ISO/IEC15288:2008, Systems and software engineering, prior to being considered. The revised INCOSE Systems Engineering Handbook v3.2 aligns with the structure and principles of ISO/IEC 15288:2008 and presents the generic SE life-cycle process steps in their entirety, without duplication or redundancy, in a single location within the text. As such, the revised Handbook v3.2 serves as a comprehensive instructional and reference manual for effectively understanding SE processes and conducting SE and better serves certification candidates preparing for the CSEP exam.

  6. Lévy matters V functionals of Lévy processes

    CERN Document Server

    Andersen, Lars Nørvang; Aurzada, Frank; Glynn, Peter W; Maejima, Makoto; Pihlsgård, Mats; Simon, Thomas

    2015-01-01

    This three-chapter volume concerns the distributions of certain functionals of Lévy processes. The first chapter, by Makoto Maejima, surveys representations of the main sub-classes of infinitesimal distributions in terms of mappings of certain Lévy processes via stochastic integration. The second chapter, by Lars Nørvang Andersen, Søren Asmussen, Peter W. Glynn and Mats Pihlsgård, concerns Lévy processes reflected at two barriers, where reflection is formulated à la Skorokhod. These processes can be used to model systems with a finite capacity, which is crucial in many real life situations, a most important quantity being the overflow or the loss occurring at the upper barrier.  If a process is killed when crossing the boundary, a natural question concerns its lifetime. Deep formulas from fluctuation theory are the key to many classical results, which are reviewed in the third chapter by Frank Aurzada and Thomas Simon. The main part, however, discusses recent advances and developments in the setting w...

  7. Implementation of the chemistry module MECCA (v2.5 in the modal aerosol version of the Community Atmosphere Model component (v3.6.33 of the Community Earth System Model

    Directory of Open Access Journals (Sweden)

    M. S. Long

    2013-02-01

    Full Text Available A coupled atmospheric chemistry and climate system model was developed using the modal aerosol version of the National Center for Atmospheric Research Community Atmosphere Model (modal-CAM; v3.6.33 and the Max Planck Institute for Chemistry's Module Efficiently Calculating the Chemistry of the Atmosphere (MECCA; v2.5 to provide enhanced resolution of multiphase processes, particularly those involving inorganic halogens, and associated impacts on atmospheric composition and climate. Three Rosenbrock solvers (Ros-2, Ros-3, RODAS-3 were tested in conjunction with the basic load-balancing options available to modal-CAM (1 to establish an optimal configuration of the implicitly-solved multiphase chemistry module that maximizes both computational speed and repeatability of Ros-2 and RODAS-3 results versus Ros-3, and (2 to identify potential implementation strategies for future versions of this and similar coupled systems. RODAS-3 was faster than Ros-2 and Ros-3 with good reproduction of Ros-3 results, while Ros-2 was both slower and substantially less reproducible relative to Ros-3 results. Modal-CAM with MECCA chemistry was a factor of 15 slower than modal-CAM using standard chemistry. MECCA chemistry integration times demonstrated a systematic frequency distribution for all three solvers, and revealed that the change in run-time performance was due to a change in the frequency distribution of chemical integration times; the peak frequency was similar for all solvers. This suggests that efficient chemistry-focused load-balancing schemes can be developed that rely on the parameters of this frequency distribution.

  8. Control-Oriented Models for SO Fuel Cells from the Angle of V&V: Analysis, Simplification Possibilities, Performance

    Directory of Open Access Journals (Sweden)

    Ekaterina Auer

    2017-12-01

    Full Text Available In this paper, we take a look at the analysis and parameter identification for control-oriented, dynamic models for the thermal subsystem of solid oxide fuel cells (SOFC from the systematized point of view of verification and validation (V&V. First, we give a possible classification of models according to their verification degree which depends, for example, on the kind of arithmetic used for both formulation and simulation. Typical SOFC models, consisting of several coupled differential equations for gas preheaters and the temperature distribution in the stack module, do not have analytical solutions because of spatial nonlinearity. Therefore, in the next part of the paper, we describe in detail two possible ways to simplify such models so that the underlying differential equations can be solved analytically while still being sufficiently accurate to serve as the basis for control synthesis. The simplifying assumption is to approximate the heat capacities of the gases by zero-order polynomials (or first-oder polynomials, respectively in the temperature. In the last, application-oriented part of the paper, we identify the parameters of these models as well as compare their performance and their ability to reflect the reality with the corresponding characteristics of models in which the heat capacities are represented by quadratic polynomials (the usual case. For this purpose, the framework UniVerMeC (Unified Framework for Verified GeoMetric Computations is used, which allows us to employ different kinds of arithmetics including the interval one. This latter possibility ensures a high level of reliability of simulations and of the subsequent validation. Besides, it helps to take into account bounded uncertainty in measurements.

  9. A Framework for Performing V&V within Reuse-Based Software Engineering

    Science.gov (United States)

    Addy, Edward A.

    1996-01-01

    Verification and validation (V&V) is performed during application development for many systems, especially safety-critical and mission-critical systems. The V&V process is intended to discover errors, especially errors related to critical processing, as early as possible during the development process. Early discovery is important in order to minimize the cost and other impacts of correcting these errors. In order to provide early detection of errors, V&V is conducted in parallel with system development, often beginning with the concept phase. In reuse-based software engineering, however, decisions on the requirements, design and even implementation of domain assets can be made prior to beginning development of a specific system. In this case, V&V must be performed during domain engineering in order to have an impact on system development. This paper describes a framework for performing V&V within architecture-centric, reuse-based software engineering. This framework includes the activities of traditional application-level V&V, and extends these activities into domain engineering and into the transition between domain engineering and application engineering. The framework includes descriptions of the types of activities to be performed during each of the life-cycle phases, and provides motivation for the activities.

  10. Modelling Hospital Materials Management Processes

    Directory of Open Access Journals (Sweden)

    Raffaele Iannone

    2013-06-01

    integrated and detailed analysis and description model for hospital materials management data and tasks, which is able to tackle information from patient requirements to usage, from replenishment requests to supplying and handling activities. The model takes account of medical risk reduction, traceability and streamlined processes perspectives. Second, the paper translates this information into a business process model and mathematical formalization.The study provides a useful guide to the various relevant technology‐related, management and business issues, laying the foundations of an efficient reengineering of the supply chain to reduce healthcare costs and improve the quality of care.

  11. The role of attention in figure-ground segregation in areas V1 and V4 of the visual cortex.

    Science.gov (United States)

    Poort, Jasper; Raudies, Florian; Wannig, Aurel; Lamme, Victor A F; Neumann, Heiko; Roelfsema, Pieter R

    2012-07-12

    Our visual system segments images into objects and background. Figure-ground segregation relies on the detection of feature discontinuities that signal boundaries between the figures and the background and on a complementary region-filling process that groups together image regions with similar features. The neuronal mechanisms for these processes are not well understood and it is unknown how they depend on visual attention. We measured neuronal activity in V1 and V4 in a task where monkeys either made an eye movement to texture-defined figures or ignored them. V1 activity predicted the timing and the direction of the saccade if the figures were task relevant. We found that boundary detection is an early process that depends little on attention, whereas region filling occurs later and is facilitated by visual attention, which acts in an object-based manner. Our findings are explained by a model with local, bottom-up computations for boundary detection and feedback processing for region filling. Copyright © 2012 Elsevier Inc. All rights reserved.

  12. Regime-switching models to study psychological process

    NARCIS (Netherlands)

    Hamaker, E.L.; Grasman, R.P.P.P.; Kamphuis, J.H.

    2010-01-01

    Many psychological processes are characterized by recurrent shifts between different states. To model these processes at the level of the individual, regime-switching models may prove useful. In this chapter we discuss two of these models: the threshold autoregressive model and the Markov

  13. Modeling biochemical transformation processes and information processing with Narrator

    Directory of Open Access Journals (Sweden)

    Palfreyman Niall M

    2007-03-01

    Full Text Available Abstract Background Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs, which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Results Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Conclusion Narrator is a

  14. Calculations of nuclear data for the reactions of neutrons and protons with heavy nuclei at energy from 1 MeV up to 2 GeV

    International Nuclear Information System (INIS)

    Konshin, V.A.

    1995-01-01

    Several nuclear model codes were applied to calculations of nuclear data in the energy region from 1 MeV to 2 GeV. At energies from 1 to 20 MeV the statistical model code STAPRE was used for calculations of the neutron cross-sections for fission, (n,2n) and (n,3n) reaction cross-sections for 71 actinide isotopes. In the energy region from 10 to 100 MeV the nuclear theory code GNASH was used to calculate the neutron fission and (n,xn) cross-sections for 238 U, 235 U, 239 Pu, 232 Th, 237 Np, 238 Pu, 241 Am, 243 Am, 245 Cm and 246 Cm. At energies from 100 MeV to 2 GeV the intranuclear cascade-exciton model including the fission process was applied to calculations of the interactions of protons and neutrons with actinides and the calculated results are compared with experimental data. (author)

  15. Probabilistic modeling of discourse-aware sentence processing.

    Science.gov (United States)

    Dubey, Amit; Keller, Frank; Sturt, Patrick

    2013-07-01

    Probabilistic models of sentence comprehension are increasingly relevant to questions concerning human language processing. However, such models are often limited to syntactic factors. This restriction is unrealistic in light of experimental results suggesting interactions between syntax and other forms of linguistic information in human sentence processing. To address this limitation, this article introduces two sentence processing models that augment a syntactic component with information about discourse co-reference. The novel combination of probabilistic syntactic components with co-reference classifiers permits them to more closely mimic human behavior than existing models. The first model uses a deep model of linguistics, based in part on probabilistic logic, allowing it to make qualitative predictions on experimental data; the second model uses shallow processing to make quantitative predictions on a broad-coverage reading-time corpus. Copyright © 2013 Cognitive Science Society, Inc.

  16. Center-of-mass corrections in the S+V potential model

    International Nuclear Information System (INIS)

    Palladino, B.E.

    1987-02-01

    Center-of-mass corrections to the mass spectrum and static properties of low-lying S-wave baryons and mesons are discussed in the context of a relativistic, independent quark model, based on a Dirac equation, with equally mixed scalar (S) and vector (V) confining potential. (author) [pt

  17. The (Mathematical) Modeling Process in Biosciences.

    Science.gov (United States)

    Torres, Nestor V; Santos, Guido

    2015-01-01

    In this communication, we introduce a general framework and discussion on the role of models and the modeling process in the field of biosciences. The objective is to sum up the common procedures during the formalization and analysis of a biological problem from the perspective of Systems Biology, which approaches the study of biological systems as a whole. We begin by presenting the definitions of (biological) system and model. Particular attention is given to the meaning of mathematical model within the context of biology. Then, we present the process of modeling and analysis of biological systems. Three stages are described in detail: conceptualization of the biological system into a model, mathematical formalization of the previous conceptual model and optimization and system management derived from the analysis of the mathematical model. All along this work the main features and shortcomings of the process are analyzed and a set of rules that could help in the task of modeling any biological system are presented. Special regard is given to the formative requirements and the interdisciplinary nature of this approach. We conclude with some general considerations on the challenges that modeling is posing to current biology.

  18. V Congress of Spanish Geochemistry

    International Nuclear Information System (INIS)

    1993-01-01

    This proceedings book present the lectures of V Spanish geochemistry Congress. The sessions were: 1.- Materials geochemistry and geologic process. 2.- Geochemistry prospection 3.- Environmental geochemistry 4.- Isotopic geochemistry 5.- Organic geochemistry 6.- Natural materials geochemistry for industry 7.- Hydrogeochemistry 8.- Mathematical models in geochemistry 9.- Analysis methods in geochemistry 10.-Training of geochemistry 11.-Cosmochemistry

  19. Concepts of Model Verification and Validation

    International Nuclear Information System (INIS)

    Thacker, B.H.; Doebling, S.W.; Hemez, F.M.; Anderson, M.C.; Pepin, J.E.; Rodriguez, E.A.

    2004-01-01

    VandV for all safety-related nuclear facility design, analyses, and operations. In fact, DNFSB 2002-1 recommends to the DOE and National Nuclear Security Administration (NNSA) that a VandV process be performed for all safety related software and analysis. Model verification and validation are the primary processes for quantifying and building credibility in numerical models. Verification is the process of determining that a model implementation accurately represents the developer's conceptual description of the model and its solution. Validation is the process of determining the degree to which a model is an accurate representation of the real world from the perspective of the intended uses of the model. Both verification and validation are processes that accumulate evidence of a model's correctness or accuracy for a specific scenario; thus, VandV cannot prove that a model is correct and accurate for all possible scenarios, but, rather, it can provide evidence that the model is sufficiently accurate for its intended use. Model VandV is fundamentally different from software VandV. Code developers developing computer programs perform software VandV to ensure code correctness, reliability, and robustness. In model VandV, the end product is a predictive model based on fundamental physics of the problem being solved. In all applications of practical interest, the calculations involved in obtaining solutions with the model require a computer code, e.g., finite element or finite difference analysis. Therefore, engineers seeking to develop credible predictive models critically need model VandV guidelines and procedures. The expected outcome of the model VandV process is the quantified level of agreement between experimental data and model prediction, as well as the predictive accuracy of the model. This report attempts to describe the general philosophy, definitions, concepts, and processes for conducting a successful VandV program. This objective is motivated by the need for

  20. Model for Simulating a Spiral Software-Development Process

    Science.gov (United States)

    Mizell, Carolyn; Curley, Charles; Nayak, Umanath

    2010-01-01

    A discrete-event simulation model, and a computer program that implements the model, have been developed as means of analyzing a spiral software-development process. This model can be tailored to specific development environments for use by software project managers in making quantitative cases for deciding among different software-development processes, courses of action, and cost estimates. A spiral process can be contrasted with a waterfall process, which is a traditional process that consists of a sequence of activities that include analysis of requirements, design, coding, testing, and support. A spiral process is an iterative process that can be regarded as a repeating modified waterfall process. Each iteration includes assessment of risk, analysis of requirements, design, coding, testing, delivery, and evaluation. A key difference between a spiral and a waterfall process is that a spiral process can accommodate changes in requirements at each iteration, whereas in a waterfall process, requirements are considered to be fixed from the beginning and, therefore, a waterfall process is not flexible enough for some projects, especially those in which requirements are not known at the beginning or may change during development. For a given project, a spiral process may cost more and take more time than does a waterfall process, but may better satisfy a customer's expectations and needs. Models for simulating various waterfall processes have been developed previously, but until now, there have been no models for simulating spiral processes. The present spiral-process-simulating model and the software that implements it were developed by extending a discrete-event simulation process model of the IEEE 12207 Software Development Process, which was built using commercially available software known as the Process Analysis Tradeoff Tool (PATT). Typical inputs to PATT models include industry-average values of product size (expressed as number of lines of code

  1. Software-Engineering Process Simulation (SEPS) model

    Science.gov (United States)

    Lin, C. Y.; Abdel-Hamid, T.; Sherif, J. S.

    1992-01-01

    The Software Engineering Process Simulation (SEPS) model is described which was developed at JPL. SEPS is a dynamic simulation model of the software project development process. It uses the feedback principles of system dynamics to simulate the dynamic interactions among various software life cycle development activities and management decision making processes. The model is designed to be a planning tool to examine tradeoffs of cost, schedule, and functionality, and to test the implications of different managerial policies on a project's outcome. Furthermore, SEPS will enable software managers to gain a better understanding of the dynamics of software project development and perform postmodern assessments.

  2. Minimal Z′ models and the 125 GeV Higgs boson

    International Nuclear Information System (INIS)

    Basso, L.

    2013-01-01

    The 1-loop renormalization group equations for the minimal Z ′ models encompassing a type-I seesaw mechanism are studied in the light of the 125 GeV Higgs boson observation. This model is taken as a benchmark for the general case of singlet extensions of the standard model. The most important result is that negative scalar mixing angles are favored with respect to positive values. Further, a minimum value for the latter exists, as well as a maximum value for the masses of the heavy neutrinos, depending on the vacuum expectation value of the singlet scalar

  3. Shallow water processes govern system-wide phytoplankton bloom dynamics: A modeling study

    Science.gov (United States)

    Lucas, L.V.; Koseff, Jeffrey R.; Monismith, Stephen G.; Thompson, J.K.

    2009-01-01

    A pseudo-two-dimensional numerical model of estuarine phytoplankton growth and consumption, vertical turbulent mixing, and idealized cross-estuary transport was developed and applied to South San Francisco Bay. This estuary has two bathymetrically distinct habitat types (deep channel, shallow shoal) and associated differences in local net rates of phytoplankton growth and consumption, as well as differences in the water column's tendency to stratify. Because many physical and biological time scales relevant to algal population dynamics decrease with decreasing depth, process rates can be especially fast in the shallow water. We used the model to explore the potential significance of hydrodynamic connectivity between a channel and shoal and whether lateral transport can allow physical or biological processes (e.g. stratification, benthic grazing, light attenuation) in one sub-region to control phytoplankton biomass and bloom development in the adjacent sub-region. Model results for South San Francisco Bay suggest that lateral transport from a productive shoal can result in phytoplankton biomass accumulation in an adjacent deep, unproductive channel. The model further suggests that turbidity and benthic grazing in the shoal can control the occurrence of a bloom system-wide; whereas, turbidity, benthic grazing, and vertical density stratification in the channel are likely to only control local bloom occurrence or modify system-wide bloom magnitude. Measurements from a related field program are generally consistent with model-derived conclusions. ?? 2008 Elsevier B.V.

  4. Metrics for Business Process Models

    Science.gov (United States)

    Mendling, Jan

    Up until now, there has been little research on why people introduce errors in real-world business process models. In a more general context, Simon [404] points to the limitations of cognitive capabilities and concludes that humans act rationally only to a certain extent. Concerning modeling errors, this argument would imply that human modelers lose track of the interrelations of large and complex models due to their limited cognitive capabilities and introduce errors that they would not insert in a small model. A recent study by Mendling et al. [275] explores in how far certain complexity metrics of business process models have the potential to serve as error determinants. The authors conclude that complexity indeed appears to have an impact on error probability. Before we can test such a hypothesis in a more general setting, we have to establish an understanding of how we can define determinants that drive error probability and how we can measure them.

  5. Approximation to the Modelling of Charge and Discharge Processes in Electrochemical Batteries by Integral Equations

    International Nuclear Information System (INIS)

    Balenzategui, J. L.

    1999-01-01

    A new way for the modelling of the charge and discharge processes in electrochemical batteries based on the use of integral equations is presented. The proposed method models the charge curves by the so called fractional or cumulative integrals of a certain objective function f(t) that must be sought. The charge figures can be easily fitted by breaking down this objective function as the addition of two different Lorentz type functions: the first one is associated to the own charge process and the second one to the overcharge process. The method allows calculating the starting voltage for overcharge as the intersection between both functions. The curve fitting of this model to different experimental charge curves, by using the Marquart algorithm, has shown very accurate results. In the case of discharge curves, two possible methods for modelling purposes are suggested, well by using the same kind of integral equations, well by the simple subtraction of an objective function f(t) from a constant value V O D. Many other aspects for the study and analysis of this method in order to improve its results in further developments are also discussed. (Author) 10 refs

  6. Les Houches 2015: Physics at TeV Colliders Standard Model Working Group Report

    Energy Technology Data Exchange (ETDEWEB)

    Andersen, J.R.; et al.

    2016-05-16

    This Report summarizes the proceedings of the 2015 Les Houches workshop on Physics at TeV Colliders. Session 1 dealt with (I) new developments relevant for high precision Standard Model calculations, (II) the new PDF4LHC parton distributions, (III) issues in the theoretical description of the production of Standard Model Higgs bosons and how to relate experimental measurements, (IV) a host of phenomenological studies essential for comparing LHC data from Run I with theoretical predictions and projections for future measurements in Run II, and (V) new developments in Monte Carlo event generators.

  7. Les Houches 2015: Physics at TeV Colliders Standard Model Working Group Report

    CERN Document Server

    Andersen, J.R.; Becker, K.; Bell, M.; Bellm, J.; Bendavid, J.; Bothmann, E.; Boughezal, R.; Butterworth, J.; Carrazza, S.; Chiesa, M.; Cieri, L.; Ciulli, V.; Denner, A.; Duehrssen-Debling, M.; Falmagne, G.; Forte, S.; Francavilla, P.; Frederix, R.; Freytsis, M.; Gao, J.; Gras, P.; Grazzini, M.; Greiner, N.; Grellscheid, D.; Heinrich, G.; Hesketh, G.; Hoche, S.; Hofer, L.; Hou, T.J.; Huss, A.; Huston, J.; Isaacson, J.; Jueid, A.; Kallweit, S.; Kar, D.; Kassabov, Z.; Konstantinides, V.; Krauss, F.; Kuttimalai, S.; Lazapoulos, A.; Lenzi, P.; Li, Y.; Lindert, J.M.; Liu, X.; Luisoni, G.; Lonnblad, L.; Maierhofer, P.; Maître, D.; Marini, A.C.; Montagna, G.; Moretti, M.; Nadolsky, P.M.; Nail, G.; Napoletano, D.; Nicrosini, O.; Oleari, C.; Pagani, D.; Pandini, C.; Perrozzi, L.; Petriello, F.; Piccinini, F.; Platzer, S.; Pogrebnyak, I.; Pozzorini, S.; Prestel, S.; Reuschle, C.; Rojo, J.; Russo, L.; Schichtel, P.; Schonherr, M.; Schumann, S.; Siodmok, A.; Skands, P.; Soper, D.; Soyez, G.; Sun, P.; Tackmann, F.J.; Tackmann, K.; Takasugi, E.; Thaler, J.; Uccirati, S.; Utku, U.; Viliani, L.; Vryonidou, E.; Wang, B.T.; Waugh, B.; Weber, M.A.; Williams, C.; Winter, J.; Xie, K.P.; Yuan, C.P.; Yuan, F.; Zapp, K.; Zaro, M.

    2016-01-01

    This Report summarizes the proceedings of the 2015 Les Houches workshop on Physics at TeV Colliders. Session 1 dealt with (I) new developments relevant for high precision Standard Model calculations, (II) the new PDF4LHC parton distributions, (III) issues in the theoretical description of the production of Standard Model Higgs bosons and how to relate experimental measurements, (IV) a host of phenomenological studies essential for comparing LHC data from Run I with theoretical predictions and projections for future measurements in Run II, and (V) new developments in Monte Carlo event generators.

  8. Revising process models through inductive learning

    NARCIS (Netherlands)

    Maggi, F.M.; Corapi, D.; Russo, A.; Lupu, E.; Visaggio, G.; Muehlen, zur M.; Su, J.

    2011-01-01

    Discovering the Business Process (BP) model underpinning existing practices through analysis of event logs, allows users to understand, analyse and modify the process. But, to be useful, the BP model must be kept in line with practice throughout its lifetime, as changes occur to the business

  9. Diagnosing differences between business process models

    NARCIS (Netherlands)

    Dijkman, R.M.; Dumas, M.; Reichert, M.; Shan, M.-C.

    2008-01-01

    This paper presents a technique to diagnose differences between business process models in the EPC notation. The diagnosis returns the exact position of a difference in the business process models and diagnoses the type of a difference, using a typology of differences developed in previous work.

  10. Mathematical modeling of photovoltaic thermal PV/T system with v-groove collector

    Science.gov (United States)

    Zohri, M.; Fudholi, A.; Ruslan, M. H.; Sopian, K.

    2017-07-01

    The use of v-groove in solar collector has a higher thermal efficiency in references. Dropping the working heat of photovoltaic panel was able to raise the electrical efficiency performance. Electrical and thermal efficiency were produced by photovoltaic thermal (PV/T) system concurrently. Mathematical modeling based on steady-state thermal analysis of PV/T system with v-groove was conducted. With matrix inversion method, the energy balance equations are explained by means of the investigative method. The comparison results show that in the PV/T system with the V-groove collector is higher temperature, thermal and electrical efficiency than other collectors.

  11. AMFIBIA: A Meta-Model for the Integration of Business Process Modelling Aspects

    DEFF Research Database (Denmark)

    Axenath, Björn; Kindler, Ekkart; Rubin, Vladimir

    2007-01-01

    AMFIBIA is a meta-model that formalises the essential aspects and concepts of business processes. Though AMFIBIA is not the first approach to formalising the aspects and concepts of business processes, it is more ambitious in the following respects: Firstly, it is independent from particular...... modelling formalisms of business processes and it is designed in such a way that any formalism for modelling some aspect of a business process can be plugged into AMFIBIA. Therefore, AMFIBIA is formalism-independent. Secondly, it is not biased toward any aspect of business processes; the different aspects...... can be considered and modelled independently of each other. Moreover, AMFIBIA is not restricted to a fixed set of aspects; new aspects of business processes can be easily integrated. Thirdly, AMFIBIA does not only name and relate the concepts of business process modelling, as it is typically done...

  12. Stability of phase transformation models for Ti-6Al-4V under cyclic thermal loading imposed during laser metal deposition

    Science.gov (United States)

    Klusemann, Benjamin; Bambach, Markus

    2018-05-01

    Processing conditions play a crucial role for the resulting microstructure and properties of the material. In particular, processing materials under non-equilibrium conditions can lead to a remarkable improvement of the final properties [1]. Additive manufacturing represents a specific process example considered in this study. Models for the prediction of residual stresses and microstructure in additive manufacturing processes, such as laser metal deposition, are being developed with huge efforts to support the development of materials and processes as well as to support process design [2-4]. Since the microstructure predicted after each heating and cooling cycle induced by the moving laser source enters the phase transformation kinetics and microstucture evolution of the subsequent heating and cooling cycle, a feed-back loop for the microstructure calculation is created. This calculation loop may become unstable so that the computed microstructure and related properties become very sensitive to small variations in the input parameters, e.g. thermal conductivity. In this paper, a model for phase transformation in Ti-6Al-4V, originally proposed by Charles Murgau et al. [5], is adopted and minimal adjusted concerning the decomposition of the martensite phase are made. This model is subsequently used to study the changes in the predictions of the different phase volume fractions during heating and cooling under the conditions of laser metal deposition with respect to slight variations in the thermal process history.

  13. Model based process-product design and analysis

    DEFF Research Database (Denmark)

    Gani, Rafiqul

    This paper gives a perspective on modelling and the important role it has within product-process design and analysis. Different modelling issues related to development and application of systematic model-based solution approaches for product-process design is discussed and the need for a hybrid...... model-based framework is highlighted. This framework should be able to manage knowledge-data, models, and associated methods and tools integrated with design work-flows and data-flows for specific product-process design problems. In particular, the framework needs to manage models of different types......, forms and complexity, together with their associated parameters. An example of a model-based system for design of chemicals based formulated products is also given....

  14. Understanding Quality in Process Modelling: Towards a Holistic Perspective

    Directory of Open Access Journals (Sweden)

    Jan Recker

    2007-09-01

    Full Text Available Quality is one of the main topics in current conceptual modelling research, as is the field of business process modelling. Yet, widely acknowledged academic contributions towards an understanding or measurement of business process model quality are limited at best. In this paper I argue that the development of methodical theories concerning the measurement or establishment of process model quality must be preceded by methodological elaborations on business process modelling. I further argue that existing epistemological foundations of process modelling are insufficient for describing all extrinsic and intrinsic traits of model quality. This in turn has led to a lack of holistic understanding of process modelling. Taking into account the inherent social and purpose-oriented character of process modelling in contemporary organizations I present a socio-pragmatic constructionist methodology of business process modelling and sketch out implications of this perspective towards an understanding of process model quality. I anticipate that, based on this research, theories can be developed that facilitate the evaluation of the ’goodness’ of a business process model.

  15. Animated-simulation modeling facilitates clinical-process costing.

    Science.gov (United States)

    Zelman, W N; Glick, N D; Blackmore, C C

    2001-09-01

    Traditionally, the finance department has assumed responsibility for assessing process costs in healthcare organizations. To enhance process-improvement efforts, however, many healthcare providers need to include clinical staff in process cost analysis. Although clinical staff often use electronic spreadsheets to model the cost of specific processes, PC-based animated-simulation tools offer two major advantages over spreadsheets: they allow clinicians to interact more easily with the costing model so that it more closely represents the process being modeled, and they represent cost output as a cost range rather than as a single cost estimate, thereby providing more useful information for decision making.

  16. How visual cognition influences process model comprehension

    NARCIS (Netherlands)

    Petrusel, Razvan; Mendling, Jan; Reijers, Hajo A.

    2017-01-01

    Process analysts and other professionals extensively use process models to analyze business processes and identify performance improvement opportunities. Therefore, it is important that such models can be easily and properly understood. Previous research has mainly focused on two types of factors

  17. Introducing uncertainty analysis of nucleation and crystal growth models in Process Analytical Technology (PAT) system design of crystallization processes.

    Science.gov (United States)

    Samad, Noor Asma Fazli Abdul; Sin, Gürkan; Gernaey, Krist V; Gani, Rafiqul

    2013-11-01

    This paper presents the application of uncertainty and sensitivity analysis as part of a systematic model-based process monitoring and control (PAT) system design framework for crystallization processes. For the uncertainty analysis, the Monte Carlo procedure is used to propagate input uncertainty, while for sensitivity analysis, global methods including the standardized regression coefficients (SRC) and Morris screening are used to identify the most significant parameters. The potassium dihydrogen phosphate (KDP) crystallization process is used as a case study, both in open-loop and closed-loop operation. In the uncertainty analysis, the impact on the predicted output of uncertain parameters related to the nucleation and the crystal growth model has been investigated for both a one- and two-dimensional crystal size distribution (CSD). The open-loop results show that the input uncertainties lead to significant uncertainties on the CSD, with appearance of a secondary peak due to secondary nucleation for both cases. The sensitivity analysis indicated that the most important parameters affecting the CSDs are nucleation order and growth order constants. In the proposed PAT system design (closed-loop), the target CSD variability was successfully reduced compared to the open-loop case, also when considering uncertainty in nucleation and crystal growth model parameters. The latter forms a strong indication of the robustness of the proposed PAT system design in achieving the target CSD and encourages its transfer to full-scale implementation. Copyright © 2013 Elsevier B.V. All rights reserved.

  18. The on-line coupled atmospheric chemistry model system MECO(n) - Part 5: Expanding the Multi-Model-Driver (MMD v2.0) for 2-way data exchange including data interpolation via GRID (v1.0)

    Science.gov (United States)

    Kerkweg, Astrid; Hofmann, Christiane; Jöckel, Patrick; Mertens, Mariano; Pante, Gregor

    2018-03-01

    As part of the Modular Earth Submodel System (MESSy), the Multi-Model-Driver (MMD v1.0) was developed to couple online the regional Consortium for Small-scale Modeling (COSMO) model into a driving model, which can be either the regional COSMO model or the global European Centre Hamburg general circulation model (ECHAM) (see Part 2 of the model documentation). The coupled system is called MECO(n), i.e., MESSy-fied ECHAM and COSMO models nested n times. In this article, which is part of the model documentation of the MECO(n) system, the second generation of MMD is introduced. MMD comprises the message-passing infrastructure required for the parallel execution (multiple programme multiple data, MPMD) of different models and the communication of the individual model instances, i.e. between the driving and the driven models. Initially, the MMD library was developed for a one-way coupling between the global chemistry-climate ECHAM/MESSy atmospheric chemistry (EMAC) model and an arbitrary number of (optionally cascaded) instances of the regional chemistry-climate model COSMO/MESSy. Thus, MMD (v1.0) provided only functions for unidirectional data transfer, i.e. from the larger-scale to the smaller-scale models.Soon, extended applications requiring data transfer from the small-scale model back to the larger-scale model became of interest. For instance, the original fields of the larger-scale model can directly be compared to the upscaled small-scale fields to analyse the improvements gained through the small-scale calculations, after the results are upscaled. Moreover, the fields originating from the two different models might be fed into the same diagnostic tool, e.g. the online calculation of the radiative forcing calculated consistently with the same radiation scheme. Last but not least, enabling the two-way data transfer between two models is the first important step on the way to a fully dynamical and chemical two-way coupling of the various model instances.In MMD (v1

  19. A general model for membrane-based separation processes

    DEFF Research Database (Denmark)

    Soni, Vipasha; Abildskov, Jens; Jonsson, Gunnar Eigil

    2009-01-01

    behaviour will play an important role. In this paper, modelling of membrane-based processes for separation of gas and liquid mixtures are considered. Two general models, one for membrane-based liquid separation processes (with phase change) and another for membrane-based gas separation are presented....... The separation processes covered are: membrane-based gas separation processes, pervaporation and various types of membrane distillation processes. The specific model for each type of membrane-based process is generated from the two general models by applying the specific system descriptions and the corresponding...

  20. APROMORE : an advanced process model repository

    NARCIS (Netherlands)

    La Rosa, M.; Reijers, H.A.; Aalst, van der W.M.P.; Dijkman, R.M.; Mendling, J.; Dumas, M.; García-Bañuelos, L.

    2011-01-01

    Business process models are becoming available in large numbers due to their widespread use in many industrial applications such as enterprise and quality engineering projects. On the one hand, this raises a challenge as to their proper management: how can it be ensured that the proper process model

  1. Multi-Boson Simulation for 13 TeV ATLAS Analyses

    CERN Document Server

    The ATLAS collaboration

    2016-01-01

    This note describes the Monte Carlo setup used by ATLAS to model multi-boson processes in 13 TeV $pp$ collisions. The baseline Monte Carlo generators are compared with each other in key kinematic distributions of the processes under study. Sample normalisation and assignment of systematics uncertainties are discussed.

  2. Multi-Boson Simulation for 13 TeV ATLAS Analyses

    CERN Document Server

    The ATLAS collaboration

    2017-01-01

    This note describes the Monte Carlo setup used by ATLAS to model multi-boson processes in 13 TeV pp collisions. The baseline Monte Carlo generators are compared with each other in key kinematic distributions of the processes under study. Sample normalization and assignment of systematic uncertainties are discussed.

  3. A model evaluation checklist for process-based environmental models

    Science.gov (United States)

    Jackson-Blake, Leah

    2015-04-01

    Mechanistic catchment-scale phosphorus models appear to perform poorly where diffuse sources dominate. The reasons for this were investigated for one commonly-applied model, the INtegrated model of CAtchment Phosphorus (INCA-P). Model output was compared to 18 months of daily water quality monitoring data in a small agricultural catchment in Scotland, and model structure, key model processes and internal model responses were examined. Although the model broadly reproduced dissolved phosphorus dynamics, it struggled with particulates. The reasons for poor performance were explored, together with ways in which improvements could be made. The process of critiquing and assessing model performance was then generalised to provide a broadly-applicable model evaluation checklist, incorporating: (1) Calibration challenges, relating to difficulties in thoroughly searching a high-dimensional parameter space and in selecting appropriate means of evaluating model performance. In this study, for example, model simplification was identified as a necessary improvement to reduce the number of parameters requiring calibration, whilst the traditionally-used Nash Sutcliffe model performance statistic was not able to discriminate between realistic and unrealistic model simulations, and alternative statistics were needed. (2) Data limitations, relating to a lack of (or uncertainty in) input data, data to constrain model parameters, data for model calibration and testing, and data to test internal model processes. In this study, model reliability could be improved by addressing all four kinds of data limitation. For example, there was insufficient surface water monitoring data for model testing against an independent dataset to that used in calibration, whilst additional monitoring of groundwater and effluent phosphorus inputs would help distinguish between alternative plausible model parameterisations. (3) Model structural inadequacies, whereby model structure may inadequately represent

  4. The evolution of process-based hydrologic models

    NARCIS (Netherlands)

    Clark, Martyn P.; Bierkens, Marc F.P.; Samaniego, Luis; Woods, Ross A.; Uijlenhoet, Remko; Bennett, Katrina E.; Pauwels, Valentijn R.N.; Cai, Xitian; Wood, Andrew W.; Peters-Lidard, Christa D.

    2017-01-01

    The diversity in hydrologic models has historically led to great controversy on the "correct" approach to process-based hydrologic modeling, with debates centered on the adequacy of process parameterizations, data limitations and uncertainty, and computational constraints on model analysis. In this

  5. Software Process Validation: Quantitatively Measuring the Correspondence of a Process to a Model

    National Research Council Canada - National Science Library

    Cook, Jonathan E; Wolf, Alexander L

    1997-01-01

    .... When process models and process executions diverge, something significant is happening. The authors have developed techniques for uncovering and measuring the discrepancies between models and executions, which they call process validation...

  6. Distillation modeling for a uranium refining process

    Energy Technology Data Exchange (ETDEWEB)

    Westphal, B.R.

    1996-03-01

    As part of the spent fuel treatment program at Argonne National Laboratory, a vacuum distillation process is being employed for the recovery of uranium following an electrorefining process. Distillation of a salt electrolyte, containing a eutectic mixture of lithium and potassium chlorides, from uranium is achieved by a simple batch operation and is termed {open_quotes}cathode processing{close_quotes}. The incremental distillation of electrolyte salt will be modeled by an equilibrium expression and on a molecular basis since the operation is conducted under moderate vacuum conditions. As processing continues, the two models will be compared and analyzed for correlation with actual operating results. Possible factors that may contribute to aberrations from the models include impurities at the vapor-liquid boundary, distillate reflux, anomalous pressure gradients, and mass transport phenomena at the evaporating surface. Ultimately, the purpose of either process model is to enable the parametric optimization of the process.

  7. Model Process Control Language

    Data.gov (United States)

    National Aeronautics and Space Administration — The MPC (Model Process Control) language enables the capture, communication and preservation of a simulation instance, with sufficient detail that it can be...

  8. Low scale composite Higgs model and 1.8 ˜2 TeV diboson excess

    Science.gov (United States)

    Bian, Ligong; Liu, Da; Shu, Jing

    2018-04-01

    We consider a simple solution to explain the recent diboson excess observed by ATLAS and CMS Collaborations in models with custodial symmetry SU(2)L × SU(2)R → SU(2)c. The SU(2)L triplet vector boson ρ with mass range of 1.8 ˜ 2 TeV would be produced through the Drell-Yan process with sizable diboson decay branching to account for the excess. The other SU(2)L × SU(2)R bidoublet axial vector boson a would cancel all deviations of electroweak obervables induced by ρ even if the SM fermions mix with some heavy vector-like (composite) fermions which couple to ρ (“nonuniversally partially composite”), therefore allows arbitrary couplings between each SM fermion and ρ. We present our model in the “General Composite Higgs” framework with SO(5) × U(1)X → SO(4) × U(1)X breaking at scale f and demand the first Weinberg sum rule and positive gauge boson form factors as the theoretical constraints. We find that our model can fit the diboson excess very well if the left-handed SM light quarks, charged leptons and tops have zero, zero/moderately small and moderate/large composite components for reasonable values of gρ and f. The correlation between tree level S parameter and the h → Zγ suggest a large a contribution to h → Zγ and it is indeed a 𝒪(1) effect in our parameter space which provides a strong hint for our scenario if this diboson excess is confirmed by the 13 ˜ 14 TeV LHC Run II.

  9. Correctness-preserving configuration of business process models

    NARCIS (Netherlands)

    Aalst, van der W.M.P.; Dumas, M.; Gottschalk, F.; Hofstede, ter A.H.M.; La Rosa, M.; Mendling, J.; Fiadeiro, J.; Inverardi, P.

    2008-01-01

    Reference process models capture recurrent business operations in a given domain such as procurement or logistics. These models are intended to be configured to fit the requirements of specific organizations or projects, leading to individualized process models that are subsequently used for domain

  10. Process models and model-data fusion in dendroecology

    Directory of Open Access Journals (Sweden)

    Joel eGuiot

    2014-08-01

    Full Text Available Dendrochronology (i.e. the study of annually dated tree-ring time series has proved to be a powerful technique to understand tree-growth. This paper intends to show the interest of using ecophysiological modeling not only to understand and predict tree-growth (dendroecology but also to reconstruct past climates (dendroclimatology. Process models have been used for several decades in dendroclimatology, but it is only recently that methods of model-data fusion have led to significant progress in modeling tree-growth as a function of climate and in reconstructing past climates. These model-data fusion (MDF methods, mainly based on the Bayesian paradigm, have been shown to be powerful for both model calibration and model inversion. After a rapid survey of tree-growth modeling, we illustrate MDF with examples based on series of Southern France Aleppo pines and Central France oaks. These examples show that if plants experienced CO2 fertilization, this would have a significant effect on tree-growth which in turn would bias the climate reconstructions. This bias could be extended to other environmental non-climatic factors directly or indirectly affecting annual ring formation and not taken into account in classical empirical models, which supports the use of more complex process-based models. Finally, we conclude by showing the interest of the data assimilation methods applied in climatology to produce climate re-analyses.

  11. Catastrophe Insurance Modeled by Shot-Noise Processes

    Directory of Open Access Journals (Sweden)

    Thorsten Schmidt

    2014-02-01

    Full Text Available Shot-noise processes generalize compound Poisson processes in the following way: a jump (the shot is followed by a decline (noise. This constitutes a useful model for insurance claims in many circumstances; claims due to natural disasters or self-exciting processes exhibit similar features. We give a general account of shot-noise processes with time-inhomogeneous drivers inspired by recent results in credit risk. Moreover, we derive a number of useful results for modeling and pricing with shot-noise processes. Besides this, we obtain some highly tractable examples and constitute a useful modeling tool for dynamic claims processes. The results can in particular be used for pricing Catastrophe Bonds (CAT bonds, a traded risk-linked security. Additionally, current results regarding the estimation of shot-noise processes are reviewed.

  12. Modeling Aspects Of Activated Sludge Processes Part I: Process Modeling Of Activated Sludge Facilitation And Sedimentation

    International Nuclear Information System (INIS)

    Ibrahim, H. I.; EI-Ahwany, A.H.; Ibrahim, G.

    2004-01-01

    Process modeling of activated sludge flocculation and sedimentation reviews consider the activated sludge floc characteristics such as: morphology viable and non-viable cell ratio density and water content, bio flocculation and its kinetics were studied considering the characteristics of bio flocculation and explaining theory of Divalent Cation Bridging which describes the major role of cations in bio flocculation. Activated sludge flocculation process modeling was studied considering mass transfer limitations from Clifft and Andrew, 1981, Benefild and Molz 1983 passing Henze 1987, until Tyagi 1996 and G. Ibrahim et aI. 2002. Models of aggregation and breakage of flocs were studied by Spicer and Pratsinis 1996,and Biggs 2002 Size distribution of floes influences mass transfer and biomass separation in the activated sludge process. Therefore, it is of primary importance to establish the role of specific process operation factors, such as sludge loading dynamic sludge age and dissolved oxygen, on this distribution with special emphasis on the formation of primary particles

  13. Cholinergic and serotonergic modulation of visual information processing in monkey V1.

    Science.gov (United States)

    Shimegi, Satoshi; Kimura, Akihiro; Sato, Akinori; Aoyama, Chisa; Mizuyama, Ryo; Tsunoda, Keisuke; Ueda, Fuyuki; Araki, Sera; Goya, Ryoma; Sato, Hiromichi

    2016-09-01

    The brain dynamically changes its input-output relationship depending on the behavioral state and context in order to optimize information processing. At the molecular level, cholinergic/monoaminergic transmitters have been extensively studied as key players for the state/context-dependent modulation of brain function. In this paper, we review how cortical visual information processing in the primary visual cortex (V1) of macaque monkey, which has a highly differentiated laminar structure, is optimized by serotonergic and cholinergic systems by examining anatomical and in vivo electrophysiological aspects to highlight their similarities and distinctions. We show that these two systems have a similar layer bias for axonal fiber innervation and receptor distribution. The common target sites are the geniculorecipient layers and geniculocortical fibers, where the appropriate gain control is established through a geniculocortical signal transformation. Both systems exert activity-dependent response gain control across layers, but in a manner consistent with the receptor subtype. The serotonergic receptors 5-HT1B and 5HT2A modulate the contrast-response curve in a manner consistent with bi-directional response gain control, where the sign (facilitation/suppression) is switched according to the firing rate and is complementary to the other. On the other hand, cholinergic nicotinic/muscarinic receptors exert mono-directional response gain control without a sign reversal. Nicotinic receptors increase the response magnitude in a multiplicative manner, while muscarinic receptors exert both suppressive and facilitative effects. We discuss the implications of the two neuromodulator systems in hierarchical visual signal processing in V1 on the basis of the developed laminar structure. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.

  14. Detailed data set of a large administration building as a validation model for DIN V18599-software; Umfangreicher Validierungsdatensatz eines grossen Verwaltungsgebaeudes fuer Software zur DIN V 18599

    Energy Technology Data Exchange (ETDEWEB)

    Hoettges, Kirsten; Woessner, Simon; de Boer, Jan; Erhorn, Hans [Fraunhofer-Institut fuer Bauphysik (IBP), Abt. Energiesysteme, Kassel (Germany); Fraunhofer-Institut fuer Bauphysik (IBP), Stuttgart (Germany)

    2009-04-15

    The calculation method of DIN V 18599 represents a very complex model for the estimation of the energy efficiency of buildings. The method is used for certificates too, thus, the number of users is quite high. This fact and the complexity of the method causes high demands on the related software products. Most of the end user software tools work with the calculation engine ibp18599kernel developed by Fraunhofer-Institut for Building Physics. There is a continuous quality control for both, the kernel and the user interfaces, i.e. the end user software. This paper gives an overview of the process of quality control as well as a documentation of a validation model used within this process, i.e. a complex administration building as a sample. (Abstract Copyright [2009], Wiley Periodicals, Inc.) [German] Mit dem Berechnungsverfahren der DIN V 18599 liegt ein umfassendes Berechnungsmodell fuer die energetische Bewertung von Gebaeuden vor. Der grosse Umfang der Verfahren stellt auch an die softwaretechnischen Umsetzungen fuer die Planungspraxis hohe Anforderungen. Das Fraunhofer-Institut fuer Bauphysik hat hierzu den Rechenkern ibp18599kernel entwickelt, der zwischenzeit-lich von zahlreichen Softwarehaeusern fuer Endanwendungsprogramme zur energetischen Bewertung und Erstellung von Energieausweisen eingesetzt wird. Dieser ''Rechenmotor'' unterliegt einer steten Qualitaetskontrolle zur Sicherstellung der Berechnungsgenauigkeit. Der vorliegende Aufsatz stellt den Prozess der Qualitaetssicherung anhand eines neuen Validierungsbeispiels in Form eines komplexen Verwaltungsgebaeudes vor. Die Fallstudie kann auch zur Validierung anderer Berechnungssysteme, die nicht auf dem Rechenkern basieren, genutzt werden. (Abstract Copyright [2009], Wiley Periodicals, Inc.)

  15. A Petri Net-Based Software Process Model for Developing Process-Oriented Information Systems

    Science.gov (United States)

    Li, Yu; Oberweis, Andreas

    Aiming at increasing flexibility, efficiency, effectiveness, and transparency of information processing and resource deployment in organizations to ensure customer satisfaction and high quality of products and services, process-oriented information systems (POIS) represent a promising realization form of computerized business information systems. Due to the complexity of POIS, explicit and specialized software process models are required to guide POIS development. In this chapter we characterize POIS with an architecture framework and present a Petri net-based software process model tailored for POIS development with consideration of organizational roles. As integrated parts of the software process model, we also introduce XML nets, a variant of high-level Petri nets as basic methodology for business processes modeling, and an XML net-based software toolset providing comprehensive functionalities for POIS development.

  16. A probabilistic evaluation procedure for process model matching techniques

    NARCIS (Netherlands)

    Kuss, Elena; Leopold, Henrik; van der Aa, Han; Stuckenschmidt, Heiner; Reijers, Hajo A.

    2018-01-01

    Process model matching refers to the automatic identification of corresponding activities between two process models. It represents the basis for many advanced process model analysis techniques such as the identification of similar process parts or process model search. A central problem is how to

  17. TeV gravity at neutrino telescopes

    CERN Document Server

    Illana, J I; Meloni, D

    2005-01-01

    Cosmogenic neutrinos reach the Earth with energies around 10^9 GeV, and their interactions with matter will be measured in upcoming experiments (Auger, IceCube). Models with extra dimensions and the fundamental scale at the TeV could imply signals in these experiments. In particular, the production of microscopic black holes by cosmogenic neutrinos has been extensively studied in the literature. Here we make a complete analysis of gravity-mediated interactions at larger distances, where they can be calculated in the eikonal approximation. In these processes a neutrino of energy E_\

  18. Orientation-Cue Invariant Population Responses to Contrast-Modulated and Phase-Reversed Contour Stimuli in Macaque V1 and V2

    Science.gov (United States)

    An, Xu; Gong, Hongliang; Yin, Jiapeng; Wang, Xiaochun; Pan, Yanxia; Zhang, Xian; Lu, Yiliang; Yang, Yupeng; Toth, Zoltan; Schiessl, Ingo; McLoughlin, Niall; Wang, Wei

    2014-01-01

    Visual scenes can be readily decomposed into a variety of oriented components, the processing of which is vital for object segregation and recognition. In primate V1 and V2, most neurons have small spatio-temporal receptive fields responding selectively to oriented luminance contours (first order), while only a subgroup of neurons signal non-luminance defined contours (second order). So how is the orientation of second-order contours represented at the population level in macaque V1 and V2? Here we compared the population responses in macaque V1 and V2 to two types of second-order contour stimuli generated either by modulation of contrast or phase reversal with those to first-order contour stimuli. Using intrinsic signal optical imaging, we found that the orientation of second-order contour stimuli was represented invariantly in the orientation columns of both macaque V1 and V2. A physiologically constrained spatio-temporal energy model of V1 and V2 neuronal populations could reproduce all the recorded population responses. These findings suggest that, at the population level, the primate early visual system processes the orientation of second-order contours initially through a linear spatio-temporal filter mechanism. Our results of population responses to different second-order contour stimuli support the idea that the orientation maps in primate V1 and V2 can be described as a spatial-temporal energy map. PMID:25188576

  19. Properties of spatial Cox process models

    DEFF Research Database (Denmark)

    Møller, Jesper

    Probabilistic properties of Cox processes of relevance for statistical modelling and inference are studied. Particularly, we study the most important classes of Cox processes, including log Gaussian Cox processes, shot noise Cox processes, and permanent Cox processes. We consider moment properties...... and point process operations such as thinning, displacements, and superpositioning. We also discuss how to simulate specific Cox processes....

  20. SModelS v1.1 user manual: Improving simplified model constraints with efficiency maps

    Science.gov (United States)

    Ambrogi, Federico; Kraml, Sabine; Kulkarni, Suchita; Laa, Ursula; Lessa, Andre; Magerl, Veronika; Sonneveld, Jory; Traub, Michael; Waltenberger, Wolfgang

    2018-06-01

    SModelS is an automatized tool for the interpretation of simplified model results from the LHC. It allows to decompose models of new physics obeying a Z2 symmetry into simplified model components, and to compare these against a large database of experimental results. The first release of SModelS, v1.0, used only cross section upper limit maps provided by the experimental collaborations. In this new release, v1.1, we extend the functionality of SModelS to efficiency maps. This increases the constraining power of the software, as efficiency maps allow to combine contributions to the same signal region from different simplified models. Other new features of version 1.1 include likelihood and χ2 calculations, extended information on the topology coverage, an extended database of experimental results as well as major speed upgrades for both the code and the database. We describe in detail the concepts and procedures used in SModelS v1.1, explaining in particular how upper limits and efficiency map results are dealt with in parallel. Detailed instructions for code usage are also provided.

  1. Model for amorphous aggregation processes

    Science.gov (United States)

    Stranks, Samuel D.; Ecroyd, Heath; van Sluyter, Steven; Waters, Elizabeth J.; Carver, John A.; von Smekal, Lorenz

    2009-11-01

    The amorphous aggregation of proteins is associated with many phenomena, ranging from the formation of protein wine haze to the development of cataract in the eye lens and the precipitation of recombinant proteins during their expression and purification. While much literature exists describing models for linear protein aggregation, such as amyloid fibril formation, there are few reports of models which address amorphous aggregation. Here, we propose a model to describe the amorphous aggregation of proteins which is also more widely applicable to other situations where a similar process occurs, such as in the formation of colloids and nanoclusters. As first applications of the model, we have tested it against experimental turbidimetry data of three proteins relevant to the wine industry and biochemistry, namely, thaumatin, a thaumatinlike protein, and α -lactalbumin. The model is very robust and describes amorphous experimental data to a high degree of accuracy. Details about the aggregation process, such as shape parameters of the aggregates and rate constants, can also be extracted.

  2. Online Rule Generation Software Process Model

    OpenAIRE

    Sudeep Marwaha; Alka Aroa; Satma M C; Rajni Jain; R C Goyal

    2013-01-01

    For production systems like expert systems, a rule generation software can facilitate the faster deployment. The software process model for rule generation using decision tree classifier refers to the various steps required to be executed for the development of a web based software model for decision rule generation. The Royce’s final waterfall model has been used in this paper to explain the software development process. The paper presents the specific output of various steps of modified wat...

  3. Flexible Software Process Lines in Practice

    DEFF Research Database (Denmark)

    Kuhrmann, Marco; Ternité, Thomas; Friedrich, Jan

    2016-01-01

    Process flexibility and adaptability is a frequently discussed topic in literature, and several approaches propose techniques to improve and optimize software processes for a given organization- or project context. A software process line (SPrL) is an instrument to systematically construct...... that can be adapted to the respective context. In this article, we present an approach to construct flexible software process lines and show its practical application in the German V-Modell XT. The presented approach emerges from a 10-year research endeavor and was used to enhance the metamodel of the V......-Modell XT and to allow for improved process variability and lifecycle management. Practical dissemination and complementing empirical research show the suitability of the concept. We therefore contribute a proven approach that is presented as metamodel fragment for reuse and implementation in further...

  4. Modeling of the loss of soil by water erosion of the basin of the River V Anniversary Cuyaguateje

    International Nuclear Information System (INIS)

    Alonso, Gustavo R.; Días, Jorge; Ruíz, Maria Elena

    2008-01-01

    The complexity of the processes involved in water erosion of soils has led to widespread use of models with high level of empiricism. However, there are few applications based on models with a considerable physical basis in this field. The purpose of this work is to evaluate the potential of a model of physical basis for estimating soil loss by erosion basin-scale and analyze the behavior of the variables in this model response. The study area was located in the Sub-basin V anniversary, which belongs to the basin of the Cuyaguateje, in the province of Pinar de Rio. You were a database of physical properties of main soils of the basin, the series-temporales of solid spending and runoff measured at River, and rain recorded by a network of rain gauges across the basin. The equation of physical basis used was the sediment transport model (STM), according to Biesemans (2000). As input variables of the model were obtained the following maps: the digital elevation model, accumulative area of drainage, drainage, land use, surface water retention capacity, retention of moisture and hydraulic conductivity of saturation curve. Soil loss was obtained per pixel, and these were correlated with each time series. The results show that the process can be extended to other sub-basins without the need to validate all the variables involved

  5. Fermentation process diagnosis using a mathematical model

    Energy Technology Data Exchange (ETDEWEB)

    Yerushalmi, L; Volesky, B; Votruba, J

    1988-09-01

    Intriguing physiology of a solvent-producing strain of Clostridium acetobutylicum led to the synthesis of a mathematical model of the acetone-butanol fermentation process. The model presented is capable of describing the process dynamics and the culture behavior during a standard and a substandard acetone-butanol fermentation. In addition to the process kinetic parameters, the model includes the culture physiological parameters, such as the cellular membrane permeability and the number of membrane sites for active transport of sugar. Computer process simulation studies for different culture conditions used the model, and quantitatively pointed out the importance of selected culture parameters that characterize the cell membrane behaviour and play an important role in the control of solvent synthesis by the cell. The theoretical predictions by the new model were confirmed by experimental determination of the cellular membrane permeability.

  6. A Generic Modeling Process to Support Functional Fault Model Development

    Science.gov (United States)

    Maul, William A.; Hemminger, Joseph A.; Oostdyk, Rebecca; Bis, Rachael A.

    2016-01-01

    Functional fault models (FFMs) are qualitative representations of a system's failure space that are used to provide a diagnostic of the modeled system. An FFM simulates the failure effect propagation paths within a system between failure modes and observation points. These models contain a significant amount of information about the system including the design, operation and off nominal behavior. The development and verification of the models can be costly in both time and resources. In addition, models depicting similar components can be distinct, both in appearance and function, when created individually, because there are numerous ways of representing the failure space within each component. Generic application of FFMs has the advantages of software code reuse: reduction of time and resources in both development and verification, and a standard set of component models from which future system models can be generated with common appearance and diagnostic performance. This paper outlines the motivation to develop a generic modeling process for FFMs at the component level and the effort to implement that process through modeling conventions and a software tool. The implementation of this generic modeling process within a fault isolation demonstration for NASA's Advanced Ground System Maintenance (AGSM) Integrated Health Management (IHM) project is presented and the impact discussed.

  7. Deterministic geologic processes and stochastic modeling

    International Nuclear Information System (INIS)

    Rautman, C.A.; Flint, A.L.

    1992-01-01

    This paper reports that recent outcrop sampling at Yucca Mountain, Nevada, has produced significant new information regarding the distribution of physical properties at the site of a potential high-level nuclear waste repository. consideration of the spatial variability indicates that her are a number of widespread deterministic geologic features at the site that have important implications for numerical modeling of such performance aspects as ground water flow and radionuclide transport. Because the geologic processes responsible for formation of Yucca Mountain are relatively well understood and operate on a more-or-less regional scale, understanding of these processes can be used in modeling the physical properties and performance of the site. Information reflecting these deterministic geologic processes may be incorporated into the modeling program explicitly using geostatistical concepts such as soft information, or implicitly, through the adoption of a particular approach to modeling

  8. Adaptive gain modulation in V1 explains contextual modifications during bisection learning.

    Directory of Open Access Journals (Sweden)

    Roland Schäfer

    2009-12-01

    Full Text Available The neuronal processing of visual stimuli in primary visual cortex (V1 can be modified by perceptual training. Training in bisection discrimination, for instance, changes the contextual interactions in V1 elicited by parallel lines. Before training, two parallel lines inhibit their individual V1-responses. After bisection training, inhibition turns into non-symmetric excitation while performing the bisection task. Yet, the receptive field of the V1 neurons evaluated by a single line does not change during task performance. We present a model of recurrent processing in V1 where the neuronal gain can be modulated by a global attentional signal. Perceptual learning mainly consists in strengthening this attentional signal, leading to a more effective gain modulation. The model reproduces both the psychophysical results on bisection learning and the modified contextual interactions observed in V1 during task performance. It makes several predictions, for instance that imagery training should improve the performance, or that a slight stimulus wiggling can strongly affect the representation in V1 while performing the task. We conclude that strengthening a top-down induced gain increase can explain perceptual learning, and that this top-down signal can modify lateral interactions within V1, without significantly changing the classical receptive field of V1 neurons.

  9. Path modeling and process control

    DEFF Research Database (Denmark)

    Høskuldsson, Agnar; Rodionova, O.; Pomerantsev, A.

    2007-01-01

    and having three or more stages. The methods are applied to a process control of a multi-stage production process having 25 variables and one output variable. When moving along the process, variables change their roles. It is shown how the methods of path modeling can be applied to estimate variables...... be performed regarding the foreseeable output property y, and with respect to an admissible range of correcting actions for the parameters of the next stage. In this paper the basic principles of path modeling is presented. The mathematics is presented for processes having only one stage, having two stages...... of the next stage with the purpose of obtaining optimal or almost optimal quality of the output variable. An important aspect of the methods presented is the possibility of extensive graphic analysis of data that can provide the engineer with a detailed view of the multi-variate variation in data....

  10. User's manual for a process model code

    International Nuclear Information System (INIS)

    Kern, E.A.; Martinez, D.P.

    1981-03-01

    The MODEL code has been developed for computer modeling of materials processing facilities associated with the nuclear fuel cycle. However, it can also be used in other modeling applications. This report provides sufficient information for a potential user to apply the code to specific process modeling problems. Several examples that demonstrate most of the capabilities of the code are provided

  11. Sato Processes in Default Modeling

    DEFF Research Database (Denmark)

    Kokholm, Thomas; Nicolato, Elisa

    In reduced form default models, the instantaneous default intensity is classically the modeling object. Survival probabilities are then given by the Laplace transform of the cumulative hazard defined as the integrated intensity process. Instead, recent literature has shown a tendency towards...

  12. [The dual process model of addiction. Towards an integrated model?].

    Science.gov (United States)

    Vandermeeren, R; Hebbrecht, M

    2012-01-01

    Neurobiology and cognitive psychology have provided us with a dual process model of addiction. According to this model, behavior is considered to be the dynamic result of a combination of automatic and controlling processes. In cases of addiction the balance between these two processes is severely disturbed. Automated processes will continue to produce impulses that ensure the continuance of addictive behavior. Weak, reflective or controlling processes are both the reason for and the result of the inability to forgo addiction. To identify features that are common to current neurocognitive insights into addiction and psychodynamic views on addiction. The picture that emerges from research is not clear. There is some evidence that attentional bias has a causal effect on addiction. There is no evidence that automatic associations have a causal effect, but there is some evidence that automatic action-tendencies do have a causal effect. Current neurocognitive views on the dual process model of addiction can be integrated with an evidence-based approach to addiction and with psychodynamic views on addiction.

  13. The on-line coupled atmospheric chemistry model system MECO(n – Part 5: Expanding the Multi-Model-Driver (MMD v2.0 for 2-way data exchange including data interpolation via GRID (v1.0

    Directory of Open Access Journals (Sweden)

    A. Kerkweg

    2018-03-01

    Full Text Available As part of the Modular Earth Submodel System (MESSy, the Multi-Model-Driver (MMD v1.0 was developed to couple online the regional Consortium for Small-scale Modeling (COSMO model into a driving model, which can be either the regional COSMO model or the global European Centre Hamburg general circulation model (ECHAM (see Part 2 of the model documentation. The coupled system is called MECO(n, i.e., MESSy-fied ECHAM and COSMO models nested n times. In this article, which is part of the model documentation of the MECO(n system, the second generation of MMD is introduced. MMD comprises the message-passing infrastructure required for the parallel execution (multiple programme multiple data, MPMD of different models and the communication of the individual model instances, i.e. between the driving and the driven models. Initially, the MMD library was developed for a one-way coupling between the global chemistry–climate ECHAM/MESSy atmospheric chemistry (EMAC model and an arbitrary number of (optionally cascaded instances of the regional chemistry–climate model COSMO/MESSy. Thus, MMD (v1.0 provided only functions for unidirectional data transfer, i.e. from the larger-scale to the smaller-scale models.Soon, extended applications requiring data transfer from the small-scale model back to the larger-scale model became of interest. For instance, the original fields of the larger-scale model can directly be compared to the upscaled small-scale fields to analyse the improvements gained through the small-scale calculations, after the results are upscaled. Moreover, the fields originating from the two different models might be fed into the same diagnostic tool, e.g. the online calculation of the radiative forcing calculated consistently with the same radiation scheme. Last but not least, enabling the two-way data transfer between two models is the first important step on the way to a fully dynamical and chemical two-way coupling of the various model

  14. Mathematical Modeling of Acclimation Processes of the Photosynthetic Chain

    Directory of Open Access Journals (Sweden)

    S Heidari

    2016-10-01

    Full Text Available Introduction Photosynthetic energy conversion efficiency is characteristic of a system which is determined by interactions between various components of the system. The complex process of photosynthesis has been studied as a whole system which enables in silico examination of a large number of candidate genes for genetic engineering for a higher photosynthetic energy conversion efficiency. One of the most important environmental factors which influence the photosynthesis efficiency is light regime which can cause producing ROS components. To acclimate to such fluctuations, plants have evolved adaptive mechanisms to minimize damage of the photosynthetic apparatus excess light. A fast compatibility response to high light stresses is non-photochemical quenching process (NPQ, dissipating excessive energy to heat. Light harvested state switches into a quenched state by a conformational change of light harvesting complex (LHCII that regulated by xanthophylls and the PsbS protein within seconds. Low lumen pH activates xanthophyll synthesis via a xanthophyll cycle which consists of the de-epoxidation of violaxanthin to zeaxanthin by violaxanthin de-epoxidase in high light and inversely by zeaxanthin epoxidase in low light which occurs more slowly. Materials and Methods Thale cress (Arabidopsis thaliana (Chlombia-0 were grown on soil at 25/22 °C day/night temperature, with a 16/8 h photoperiod, and 40-70% (depend of plant species relative humidity. The light intensity was 150–200 µE m-2s-1 white light. Intensity of chlorophyll fluorescence was measured with PAM-2000 fluorometer (Heinz Walz, Germany and the manufacturer’s software (PamWin v.2. Results and Discussion In the present study, a dynamic kinetics amplified mathematical model was developed based on differential equations in order to predict short-term changes in NPQ in the process of adaptation to different light conditions. We investigated the stationary and dynamic behavior of the model

  15. Biological studies in animal models using [99mTc](CO)3 recombinant annexin V as diagnostic agent of apoptotic processes

    International Nuclear Information System (INIS)

    Teran, Mariella Adriana; Martinez, Elena; Reyes, Ana L.; Paolino, Andrea; Vital, Marcelo; Esperon, Patricia; Pacheco, Jose P.; Savio, Eduardo

    2011-01-01

    Introduction: There are many diseases associated with variations in the expression of apoptosis such as organ rejection after transplantation, myocardial ischemia or infarct and neurodegenerative diseases. For this reason, the early visualization of this process is relevant to set fast and effective therapeutic strategies. Methods: The precursor was prepared according to the procedure reported by R. Alberto, R. Schibli, P. Schubiger, U. Abram, and T. Kaden [Reactions with the technetium and rhenium carbonyl complexes (NEt 4 )[MX 3 (CO) 3 ]. Synthesis and structure of Tc(CN-But) 3 (CO) 3 ](NO 3 ) and (Net 4 )[Tc 2 (μ-SCH 2 CH 2 OH) 3 (CO) 3 ], Polyhedron 1996;15: 1079-89]. Recombinant annexin V was incubated with [ 99m Tc](H 2 O)3(CO) 3 + solution, previously neutralized with buffer. Biodistribution studies were performed in 8-week-old female Wistar rats. Animals were housed and treated in compliance with institutional guidelines related to animal experimentation. Work protocol was previously approved by the Animal Ethics Committee of the university. Two groups of rats were defined. One was used as control and the other group was previously injected with 150 mg/kg ip of cyclophosphamide to induce apoptosis. Results: The synthesis of carbonyl precursor achieved yields higher than 90%, and the radiolabeled protein was obtained with 92% of radiochemical purity and high stability in vitro. An important uptake in apoptotic tissues was confirmed by biodistributions, scintigraphic images and histological studies. Conclusions: Biodistribution studies revealed hepatobiliary elimination, high stability in vivo and important uptake in the reticuloendothelial system. In the pathologic model, higher uptake values correspond to the liver, spleen, lungs and femur. Histological studies confirmed the development of apoptosis at 8 and 24 h postinduction in the spleen and lymphocyte bulks in the peribronchial area. Scintigraphic images confirmed high uptake both the spleen and the

  16. Artificial intelligence versus statistical modeling and optimization of continuous bead milling process for bacterial cell lysis

    Directory of Open Access Journals (Sweden)

    Shafiul Haque

    2016-11-01

    Full Text Available AbstractFor a commercially viable recombinant intracellular protein production process, efficient cell lysis and protein release is a major bottleneck. The recovery of recombinant protein, cholesterol oxidase (COD was studied in a continuous bead milling process. A full factorial Response Surface Model (RSM design was employed and compared to Artificial Neural Networks coupled with Genetic Algorithm (ANN-GA. Significant process variables, cell slurry feed rate (A, bead load (B, cell load (C and run time (D, were investigated and optimized for maximizing COD recovery. RSM predicted an optimum of feed rate of 310.73 mL/h, bead loading of 79.9% (v/v, cell loading OD600 nm of 74, and run time of 29.9 min with a recovery of ~3.2 g/L. ANN coupled with GA predicted a maximum COD recovery of ~3.5 g/L at an optimum feed rate (mL/h: 258.08, bead loading (%, v/v: 80%, cell loading (OD600 nm: 73.99, and run time of 32 min. An overall 3.7-fold increase in productivity is obtained when compared to a batch process. Optimization and comparison of statistical vs. artificial intelligence techniques in continuous bead milling process has been attempted for the very first time in our study. We were able to successfully represent the complex non-linear multivariable dependence of enzyme recovery on bead milling parameters. The quadratic second order response functions are not flexible enough to represent such complex non-linear dependence. ANN being a summation function of multiple layers are capable to represent complex non-linear dependence of variables in this case; enzyme recovery as a function of bead milling parameters. Since GA can even optimize discontinuous functions present study cites a perfect example of using machine learning (ANN in combination with evolutionary optimization (GA for representing undefined biological functions which is the case for common industrial processes involving biological moieties.

  17. The short-circuit test results of 6.9 kV/2.3 kV 400 kVA-class YBCO model transformer with fault current limiting function

    International Nuclear Information System (INIS)

    Tomioka, A.; Bohno, T.; Kakami, S.; Isozaki, M.; Watanabe, K.; Toyama, K.; Sugiyama, S.; Konno, M.; Gosho, Y.; Okamoto, H.; Hayashi, H.; Tsutsumi, T.; Iwakuma, M.; Saito, T.; Tanabe, K.; Shiohara, Y.

    2013-01-01

    Highlights: ► We manufactured the 400 kV A-class YBCO model transformer with FCL function. ► Short-circuit test was performed by applying 6.9 kV on primary side. ► The short-circuit current was limited to 174 A for a prospective current of 559 A. ► It agreed with the design and we also confirmed the I c did not degrade. ► The results suggest the possibility to design YBCO transformers with FCL function. -- Abstract: We are developing an elemental technology for 66/6.9 kV 20 MVA-class superconducting power transformer with fault current limiting function. In order to obtain the characteristics of YBCO conductor when the AC over current supplied to the conductor, the model coils were manufactured with YBCO tapes and tested. Based on these results, we manufactured the 6.9 kV/2.3 kV 400 kVA-class YBCO model transformer with fault current limiting function and performed short-circuit test. At the 0.25 s after short-circuit, the short-circuit current of primary winding was limited to about 174 A for a prospective current of 559 A. It was consistent with the design. The I–V characteristics of the winding did not change before and after the test. We consider the model transformer to be able to withstand AC over-current with the function of current limiting. The results suggest the possibility to design YBCO superconducting transformers with fault current limiting function for practical power grid

  18. Distillation modeling for a uranium refining process

    International Nuclear Information System (INIS)

    Westphal, B.R.

    1996-01-01

    As part of the spent fuel treatment program at Argonne National Laboratory, a vacuum distillation process is being employed for the recovery of uranium following an electrorefining process. Distillation of a salt electrolyte, containing a eutectic mixture of lithium and potassium chlorides, from uranium is achieved by a simple batch operation and is termed open-quotes cathode processingclose quotes. The incremental distillation of electrolyte salt will be modeled by an equilibrium expression and on a molecular basis since the operation is conducted under moderate vacuum conditions. As processing continues, the two models will be compared and analyzed for correlation with actual operating results. Possible factors that may contribute to aberrations from the models include impurities at the vapor-liquid boundary, distillate reflux, anomalous pressure gradients, and mass transport phenomena at the evaporating surface. Ultimately, the purpose of either process model is to enable the parametric optimization of the process

  19. Specification of e-business process model for PayPal online payment process using Reo

    NARCIS (Netherlands)

    M. Xie

    2005-01-01

    textabstractE-business process modeling allows business analysts to better understand and analyze the business processes, and eventually to use software systems to automate (parts of) these business processes to achieve higher profit. To support e-business process modeling, many business process

  20. Simultaneous enhancement of phenolic compound degradations by Acinetobacter strain V2 via a step-wise continuous acclimation process.

    Science.gov (United States)

    Lin, Johnson; Sharma, Vikas; Milase, Ridwaan; Mbhense, Ntuthuko

    2016-06-01

    Phenol degradation enhancement of Acinetobacter strain V2 by a step-wise continuous acclimation process was investigated. At the end of 8 months, three stable adapted strains, designated as R, G, and Y, were developed with the sub-lethal concentration of phenol at 800, 1100, and 1400 mg/L, respectively, from 400 mg/L of V2 parent strain. All strains degraded phenol at their sub-lethal level within 24 h, their growth rate increased as the acclimation process continued and retained their degradation properties even after storing at -80 °C for more than 3 years. All adapted strains appeared coccoid with an ungranulated surface under electron microscope compared to typical rod-shaped parental strain V2 . The adapted Y strain also possessed superior degradation ability against aniline, benzoate, and toluene. This study demonstrated the use of long term acclimation process to develop efficient and better pollutant degrading bacterial strains with potentials in industrial and environmental bioremediation. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  1. Roll-to-Roll Processing of Inverted Polymer Solar Cells using Hydrated Vanadium(V)Oxide as a PEDOT:PSS Replacement.

    Science.gov (United States)

    Espinosa, Nieves; Dam, Henrik Friis; Tanenbaum, David M; Andreasen, Jens W; Jørgensen, Mikkel; Krebs, Frederik C

    2011-01-11

    The use of hydrated vanadium(V)oxide as a replacement of the commonly employed hole transporting material PEDOT:PSS was explored in this work. Polymer solar cells were prepared by spin coating on glass. Polymer solar cells and modules comprising 16 serially connected cells were prepared using full roll-to-roll (R2R) processing of all layers. The devices were prepared on flexible polyethyleneterphthalate (PET) and had the structure PET/ITO/ZnO/P3HT:PCBM/V₂O₅·(H₂O) n /Ag. The ITO and silver electrodes were processed and patterned by use of screen printing. The zinc oxide, P3HT:PCBM and vanadium(V)oxide layers were processed by slot-die coating. The hydrated vanadium(V)oxide layer was slot-die coated using an isopropanol solution of vanadyl-triisopropoxide (VTIP). Coating experiments were carried out to establish the critical thickness of the hydrated vanadium(V)oxide layer by varying the concentration of the VTIP precursor over two orders of magnitude. Hydrated vanadium(V)oxide layers were characterized by profilometry, scanning electron microscopy, energy dispersive X-ray spectroscopy, and grazing incidence wide angle X-ray scattering. The power conversion efficiency (PCE) for completed modules was up to 0.18%, in contrast to single cells where efficiencies of 0.4% were achieved. Stability tests under indoor and outdoor conditions were accomplished over three weeks on a solar tracker.

  2. V0 particle production studies at LHCb

    CERN Multimedia

    Knecht, M

    2009-01-01

    Although QCD is firmly established as the fundamental theory of strong interactions, the fragmentation process from partons into hadrons is still poorly understood. Phenomenological models tuned to Tevatron data show significant differences when extrapolated to LHC energies. The hadronization process can be probed at the LHC by studying V0 production, i.e. the production of KS mesons and Lambda hyperons. The LHCb experiment, with a rapidity range complementary to that of the other LHC detectors, offers a particularly interesting environment, covering the forward region where the existing models are very tunable but lack predictive power. The first 100 millions minimum bias events at LHCb will already provide a high-statistics and high-purity V0 sample. Measurements will include differential cross sections and production ratios for different strange particles as a function of rapidity and transverse momentum. The analysis can naturally be extended to cover heavier hyperons as well, and eventually lead, w...

  3. RF System Modelling for the JLab 12 GeV Upgrade and RIA

    International Nuclear Information System (INIS)

    Alicia Hofler; Jean Delayen; Hovater, J.; Stefan Simrock

    2003-01-01

    Jefferson Lab is using the MATLAB/Simulink library for RF systems developed for TTF as a tool to develop a model of its 12 GeV upgrade and the Rare Isotope Accelerator (RIA) to study the behavior and performance of the RF control system. The library includes elements describing a superconducting cavity with mechanical modes excited by Lorentz Force effects and a klystron including saturation characteristics. It can be applied to gradient and phase or in-phase and quadrature control for cavities operating in either a self-excited loop or generator driven mode. We will provide an overview of the theory behind the library components and present initial modeling results for Jefferson Lab's 12 GeV Upgrade and the RIA systems

  4. Sato Processes in Default Modeling

    DEFF Research Database (Denmark)

    Kokholm, Thomas; Nicolato, Elisa

    2010-01-01

    In reduced form default models, the instantaneous default intensity is the classical modeling object. Survival probabilities are then given by the Laplace transform of the cumulative hazard defined as the integrated intensity process. Instead, recent literature tends to specify the cumulative haz...

  5. Dimensional modeling: beyond data processing constraints.

    Science.gov (United States)

    Bunardzic, A

    1995-01-01

    The focus of information processing requirements is shifting from the on-line transaction processing (OLTP) issues to the on-line analytical processing (OLAP) issues. While the former serves to ensure the feasibility of the real-time on-line transaction processing (which has already exceeded a level of up to 1,000 transactions per second under normal conditions), the latter aims at enabling more sophisticated analytical manipulation of data. The OLTP requirements, or how to efficiently get data into the system, have been solved by applying the Relational theory in the form of Entity-Relation model. There is presently no theory related to OLAP that would resolve the analytical processing requirements as efficiently as Relational theory provided for the transaction processing. The "relational dogma" also provides the mathematical foundation for the Centralized Data Processing paradigm in which mission-critical information is incorporated as 'one and only one instance' of data, thus ensuring data integrity. In such surroundings, the information that supports business analysis and decision support activities is obtained by running predefined reports and queries that are provided by the IS department. In today's intensified competitive climate, businesses are finding that this traditional approach is not good enough. The only way to stay on top of things, and to survive and prosper, is to decentralize the IS services. The newly emerging Distributed Data Processing, with its increased emphasis on empowering the end user, does not seem to find enough merit in the relational database model to justify relying upon it. Relational theory proved too rigid and complex to accommodate the analytical processing needs. In order to satisfy the OLAP requirements, or how to efficiently get the data out of the system, different models, metaphors, and theories have been devised. All of them are pointing to the need for simplifying the highly non-intuitive mathematical constraints found

  6. Phenomenological optical model for p-/sup 4/He elastic scattering. [560 to 1730 MeV: Dirac equation optical model analysis

    Energy Technology Data Exchange (ETDEWEB)

    Mercer, R L [International Business Machines Corp., Yorktown Heights, N.Y. (USA); Arnold, L G; Clark, B C [Ohio State Univ., Columbus (USA). Dept. of Physics

    1978-01-30

    The results of a Dirac equation optical model analysis of p-/sup 4/He elastic scattering data are reported. The optical potential obtained at 1029 MeV reproduces the systematics of p-/sup 4/He data over the energy range from 560 to 1730 MeV.

  7. Model-based processing for underwater acoustic arrays

    CERN Document Server

    Sullivan, Edmund J

    2015-01-01

    This monograph presents a unified approach to model-based processing for underwater acoustic arrays. The use of physical models in passive array processing is not a new idea, but it has been used on a case-by-case basis, and as such, lacks any unifying structure. This work views all such processing methods as estimation procedures, which then can be unified by treating them all as a form of joint estimation based on a Kalman-type recursive processor, which can be recursive either in space or time, depending on the application. This is done for three reasons. First, the Kalman filter provides a natural framework for the inclusion of physical models in a processing scheme. Second, it allows poorly known model parameters to be jointly estimated along with the quantities of interest. This is important, since in certain areas of array processing already in use, such as those based on matched-field processing, the so-called mismatch problem either degrades performance or, indeed, prevents any solution at all. Third...

  8. Neutrino jets from high-mass WR gauge bosons in TeV-scale left-right symmetric models

    Science.gov (United States)

    Mitra, Manimala; Ruiz, Richard; Scott, Darren J.; Spannowsky, Michael

    2016-11-01

    We reexamine the discovery potential at hadron colliders of high-mass right-handed (RH) gauge bosons WR—an inherent ingredient of left-right symmetric models (LRSM). We focus on the regime where the WR is very heavy compared to the heavy Majorana neutrino N , and we investigate an alternative signature for WR→N decays. The produced neutrinos are highly boosted in this mass regime. Subsequently, their decays via off-shell WR bosons to jets, i.e., N →ℓ±jj, are highly collimated, forming a single neutrino jet (jN). The final-state collider signature is then ℓ±jN, instead of the widely studied ℓ±ℓ±j j . Present search strategies are not sensitive to this hierarchical mass regime due to the breakdown of the collider signature definition. We take into account QCD corrections beyond next-to-leading order (NLO) that are important for high-mass Drell-Yan processes at the 13 TeV Large Hadron Collider (LHC). For the first time, we evaluate WR production at NLO with threshold resummation at next-to-next-to-leading logarithm (NNLL) matched to the threshold-improved parton distributions. With these improvements, we find that a WR of mass MWR=3 (4 )[5 ] TeV and mass ratio of (mN/MWR)discovered with a 5 - 6 σ statistical significance at 13 TeV after 10 (100 )[2000 ] fb-1 of data. Extending the analysis to the hypothetical 100 TeV Very Large Hadron Collider (VLHC), 5 σ can be obtained for WR masses up to MW R=15 (30 ) with approximately 100 fb-1 (10 ab-1 ). Conversely, with 0.9 (10 )[150 ] fb-1 of 13 TeV data, MWR<3 (4 )[5 ] TeV and (mN/MWR)<0.1 can be excluded at 95% C.L.; with 100 fb-1 (2.5 ab-1 ) of 100 TeV data, MW R<22 (33 ) TeV can be excluded.

  9. Neuroscientific Model of Motivational Process

    Science.gov (United States)

    Kim, Sung-il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Reward-driven approach is the process in which motivation is generated by reward anticipation and selective approach behaviors toward reward. This process recruits the ventral striatum (reward area) in which basic stimulus-action association is formed, and is classified as an automatic motivation to which relatively less attention is assigned. By contrast, value-based decision-making is the process of evaluating various outcomes of actions, learning through positive prediction error, and calculating the value continuously. The striatum and the orbitofrontal cortex (valuation area) play crucial roles in sustaining motivation. Lastly, the goal-directed control is the process of regulating motivation through cognitive control to achieve goals. This consciously controlled motivation is associated with higher-level cognitive functions such as planning, retaining the goal, monitoring the performance, and regulating action. The anterior cingulate cortex (attention area) and the dorsolateral prefrontal cortex (cognitive control area) are the main neural circuits related to regulation of motivation. These three sub processes interact with each other by sending reward prediction error signals through dopaminergic pathway from the striatum and to the prefrontal cortex. The neuroscientific model of motivational process suggests several educational implications with regard to the generation, maintenance, and regulation of motivation to learn in the learning environment. PMID:23459598

  10. Neuroscientific model of motivational process.

    Science.gov (United States)

    Kim, Sung-Il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Reward-driven approach is the process in which motivation is generated by reward anticipation and selective approach behaviors toward reward. This process recruits the ventral striatum (reward area) in which basic stimulus-action association is formed, and is classified as an automatic motivation to which relatively less attention is assigned. By contrast, value-based decision-making is the process of evaluating various outcomes of actions, learning through positive prediction error, and calculating the value continuously. The striatum and the orbitofrontal cortex (valuation area) play crucial roles in sustaining motivation. Lastly, the goal-directed control is the process of regulating motivation through cognitive control to achieve goals. This consciously controlled motivation is associated with higher-level cognitive functions such as planning, retaining the goal, monitoring the performance, and regulating action. The anterior cingulate cortex (attention area) and the dorsolateral prefrontal cortex (cognitive control area) are the main neural circuits related to regulation of motivation. These three sub processes interact with each other by sending reward prediction error signals through dopaminergic pathway from the striatum and to the prefrontal cortex. The neuroscientific model of motivational process suggests several educational implications with regard to the generation, maintenance, and regulation of motivation to learn in the learning environment.

  11. Internal model of gravity influences configural body processing.

    Science.gov (United States)

    Barra, Julien; Senot, Patrice; Auclair, Laurent

    2017-01-01

    Human bodies are processed by a configural processing mechanism. Evidence supporting this claim is the body inversion effect, in which inversion impairs recognition of bodies more than other objects. Biomechanical configuration, as well as both visual and embodied expertise, has been demonstrated to play an important role in this effect. Nevertheless, the important factor of body inversion effect may also be linked to gravity orientation since gravity is one of the most fundamental constraints of our biology, behavior, and perception on Earth. The visual presentation of an inverted body in a typical body inversion paradigm turns the observed body upside down but also inverts the implicit direction of visual gravity in the scene. The orientation of visual gravity is then in conflict with the direction of actual gravity and may influence configural processing. To test this hypothesis, we dissociated the orientations of the body and of visual gravity by manipulating body posture. In a pretest we showed that it was possible to turn an avatar upside down (inversion relative to retinal coordinates) without inverting the orientation of visual gravity when the avatar stands on his/her hands. We compared the inversion effect in typical conditions (with gravity conflict when the avatar is upside down) to the inversion effect in conditions with no conflict between visual and physical gravity. The results of our experiment revealed that the inversion effect, as measured by both error rate and reaction time, was strongly reduced when there was no gravity conflict. Our results suggest that when an observed body is upside down (inversion relative to participants' retinal coordinates) but the orientation of visual gravity is not, configural processing of bodies might still be possible. In this paper, we discuss the implications of an internal model of gravity in the configural processing of observed bodies. Copyright © 2016 Elsevier B.V. All rights reserved.

  12. Aerospace Materials Process Modelling

    Science.gov (United States)

    1988-08-01

    Cooling Transformation diagram ( CCT diagram ) When a IT diagram is used in the heat process modelling, we suppose that a sudden cooling (instantaneous...processes. CE, chooses instead to study thermo-mechanical properties referring to a CCT diagram . This is thinked to be more reliable to give a true...k , mm-_____sml l ml A I 1 III 12.4 This determination is however based on the following approximations: i) A CCT diagram is valid only for the

  13. An enhanced close-in phase noise LC-VCO using parasitic V-NPN transistors in a CMOS process

    International Nuclear Information System (INIS)

    Gao Peijun; Min Hao; Oh, N J

    2009-01-01

    A differential LC voltage controlled oscillator (VCO) employing parasitic vertical-NPN (V-NPN) transistors as a negative g m -cell is presented to improve the close-in phase noise. The V-NPN transistors have lower flicker noise compared to MOS transistors. DC and AC characteristics of the V-NPN transistors are measured to facilitate the VCO design. The proposed VCO is implemented in a 0.18 μm CMOS RF/mixed signal process, and the measurement results show the close-in phase noise is improved by 3.5-9.1 dB from 100 Hz to 10 kHz offset compared to that of a similar CMOS VCO. The proposed VCO consumes only 0.41 mA from a 1.5 V power supply. (semiconductor integrated circuits)

  14. Advanced oxidation processes: overall models

    Energy Technology Data Exchange (ETDEWEB)

    Rodriguez, M. [Univ. de los Andes, Escuela Basica de Ingenieria, La Hechicera, Merida (Venezuela); Curco, D.; Addardak, A.; Gimenez, J.; Esplugas, S. [Dept. de Ingenieria Quimica. Univ. de Barcelona, Barcelona (Spain)

    2003-07-01

    Modelling AOPs implies to consider all the steps included in the process, that means, mass transfer, kinetic (reaction) and luminic steps. In this way, recent works develop models which relate the global reaction rate to catalyst concentration and radiation absorption. However, the application of such models requires to know what is the controlling step for the overall process. In this paper, a simple method is explained which allows to determine the controlling step. Thus, it is assumed that reactor is divided in two hypothetical zones (dark and illuminated), and according to the experimental results, obtained by varying only the reaction volume, it can be decided if reaction occurs only in the illuminated zone or in the all reactor, including dark zone. The photocatalytic degradation of phenol, by using titania degussa P-25 as catalyst, is studied as reaction model. The preliminary results obtained are presented here, showing that it seems that, in this case, reaction only occurs in the illuminated zone of photoreactor. A model is developed to explain this behaviour. (orig.)

  15. Modeling Business Processes in Public Administration

    Science.gov (United States)

    Repa, Vaclav

    During more than 10 years of its existence business process modeling became a regular part of organization management practice. It is mostly regarded. as a part of information system development or even as a way to implement some supporting technology (for instance workflow system). Although I do not agree with such reduction of the real meaning of a business process, it is necessary to admit that information technologies play an essential role in business processes (see [1] for more information), Consequently, an information system is inseparable from a business process itself because it is a cornerstone of the general basic infrastructure of a business. This fact impacts on all dimensions of business process management. One of these dimensions is the methodology that postulates that the information systems development provide the business process management with exact methods and tools for modeling business processes. Also the methodology underlying the approach presented in this paper has its roots in the information systems development methodology.

  16. Computerized simulation of YAG pulse laser welding of titanium alloy (TA6V): experimental characterization and modelling of the thermomechanical aspects of this process; Simulation numerique du soudage du TA6V par laser YAG impulsionnel: caracterisation experimentale et modelisation des aspects thermomecanique associees a ce procede

    Energy Technology Data Exchange (ETDEWEB)

    Robert, Y

    2007-09-15

    This work is a part of study which goal is to realize a computer modelling of the thermomechanical phenomena occurring during the YAG pulse laser welding of titanium alloy (TA6V). The filet welding has different heterogeneities (microstructural and mechanical). In fact, the temperature causes microstructural changes (phase transformations, precipitations) and modifies the mechanical properties. Thermomechanical modelling has thus to be established for the welding of TA6V. (author)

  17. Ultrasonic wave propagation in real-life austenitic V-butt welds: Numerical modeling and validation

    International Nuclear Information System (INIS)

    Hannemann, R.; Marklein, R.; Langenberg, K. J.; Schurig, C.; Koehler, B.; Walte, F.

    2000-01-01

    In nondestructive testing the evaluation of austenitic steel welds with ultrasound is a commonly used method. But, since the wave propagation, scattering, and diffraction effects in such complicated media are hardly understood, computer simulations are very helpful to increase the knowledge of the physical phenomena in such samples. A particularly powerful numerical time domain modeling tool is the well established Elastodynamic Finite Integration Technique (EFIT). Recently, EFIT has been extended to simulate elastic waves in inhomogeneous anisotropic media. In this paper, the step-by-step evaluation of ultrasonic wave propagation in inhomogeneous anisotropic media will be described and the results will be validated against measurements. As a simplified model, a V-butt weld with perpendicular grain structure is investigated. The coincidence between the B Scans of the simulation and the measurement of an idealized V-butt weld is remarkable and even effects predicted by theory and simulation - the appearance of two coupled quasi-SV waves - can be observed. As a next step, an improved and more realistic model of the grain orientation inside the V-butt weld is introduced. This model has been implemented in the EFIT code and has been validated against measurements. For this verification, measured and simulated B-Scans for a real-life V-butt weld have been compared and a significant coincidence has been observed. Furthermore, the main pulses in the B-Scans are interpreted by analyzing the snapshot-movies of the wavefronts

  18. The model of recombination process in TlBr

    International Nuclear Information System (INIS)

    Grigorjeva, L.; Millers, D.

    2002-01-01

    The time-resolved luminescence was used as a tool in the study of recombination process in several undoped TlBr crystals. The spectra and decay kinetics observed under electron beam excitation were investigated. Observation of several luminescence bands with different decay rates shows that more than one recombination center is involved and the recombination process is quite complicated. The band at ∼2.5 eV is dominant under 10 ns excitation pulse (electron beam or nitrogen laser pulses). The results of short-lived absorption and luminescence are used for analysis of possible mechanisms of recombination processes in TlBr

  19. Group Contribution Based Process Flowsheet Synthesis, Design and Modelling

    DEFF Research Database (Denmark)

    d'Anterroches, Loïc; Gani, Rafiqul

    2004-01-01

    This paper presents a process-group-contribution Method to model. simulate and synthesize a flowsheet. The process-group based representation of a flowsheet together with a process "property" model are presented. The process-group based synthesis method is developed on the basis of the computer...... aided molecular design methods and gives the ability to screen numerous process alternatives without the need to use the rigorous process simulation models. The process "property" model calculates the design targets for the generated flowsheet alternatives while a reverse modelling method (also...... developed) determines the design variables matching the target. A simple illustrative example highlighting the main features of the methodology is also presented....

  20. KENO3D, Visualisation Tool for KENO V.A and KENO-VI Geometry Models

    International Nuclear Information System (INIS)

    2008-01-01

    1 - Description of program or function: The KENO3D Visualization Tool for KENO Geometry Models is a powerful state-of-the-art visualization tool that enables KENO V.a users and KENO-VI to interactively display their three-dimensional geometry models. The KENO3D interactive options include: - Shaded or wire-frame images ; - Standard views such as top view, side view, front view, and isometric(3-D) view; - Rotating the model ; - Zooming in on selected locations ; - Selecting parts of the model to display ; - Editing colors and displaying legends ; - Displaying properties of any unit in the model ; - Creating cut-away views ; - Removing units from the model; - Printing image or saving image to a common graphics formats. KENO3D was developed for use by criticality safety specialists that use the KENO three-dimensional Monte Carlo criticality computer code. KENO V.a and KENO-VI are part of the SCALE (Standardized Computer Analyses for Licensing Evaluations) computer software system developed at Ridge National Laboratory (ORNL) that is widely used and accepted around the world for criticality safety analyses. 2 - Methods: KENO3D reads CSAS, KENO V.a, or KENO-VI input files. It attempts to verify that the KENO geometry input is 'legal', i.e., it conforms to the code input guidelines. KENO3D prints a warning message for illegal geometry input, and if possible, it displays the illegal KENO geometry to facilitate debugging of the input. Problems with more than 300,000 KENO V.a bodies have been successfully tested and displayed. KENO3D has the look and feel of a typical PC Windows application. Toolbar buttons are included for all major menu options. There is a setup dialog that allows the user to specify toolbars that should be displayed

  1. Software Engineering Laboratory (SEL) cleanroom process model

    Science.gov (United States)

    Green, Scott; Basili, Victor; Godfrey, Sally; Mcgarry, Frank; Pajerski, Rose; Waligora, Sharon

    1991-01-01

    The Software Engineering Laboratory (SEL) cleanroom process model is described. The term 'cleanroom' originates in the integrated circuit (IC) production process, where IC's are assembled in dust free 'clean rooms' to prevent the destructive effects of dust. When applying the clean room methodology to the development of software systems, the primary focus is on software defect prevention rather than defect removal. The model is based on data and analysis from previous cleanroom efforts within the SEL and is tailored to serve as a guideline in applying the methodology to future production software efforts. The phases that are part of the process model life cycle from the delivery of requirements to the start of acceptance testing are described. For each defined phase, a set of specific activities is discussed, and the appropriate data flow is described. Pertinent managerial issues, key similarities and differences between the SEL's cleanroom process model and the standard development approach used on SEL projects, and significant lessons learned from prior cleanroom projects are presented. It is intended that the process model described here will be further tailored as additional SEL cleanroom projects are analyzed.

  2. Business Model Process Configurations

    DEFF Research Database (Denmark)

    Taran, Yariv; Nielsen, Christian; Thomsen, Peter

    2015-01-01

    , by developing (inductively) an ontological classification framework, in view of the BM process configurations typology developed. Design/methodology/approach – Given the inconsistencies found in the business model studies (e.g. definitions, configurations, classifications) we adopted the analytical induction...

  3. Product/Process (P/P) Models For The Defense Waste Processing Facility (DWPF): Model Ranges And Validation Ranges For Future Processing

    Energy Technology Data Exchange (ETDEWEB)

    Jantzen, C. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL); Edwards, T. [Savannah River Site (SRS), Aiken, SC (United States). Savannah River National Lab. (SRNL)

    2015-09-25

    Radioactive high level waste (HLW) at the Savannah River Site (SRS) has successfully been vitrified into borosilicate glass in the Defense Waste Processing Facility (DWPF) since 1996. Vitrification requires stringent product/process (P/P) constraints since the glass cannot be reworked once it is poured into ten foot tall by two foot diameter canisters. A unique “feed forward” statistical process control (SPC) was developed for this control rather than statistical quality control (SQC). In SPC, the feed composition to the DWPF melter is controlled prior to vitrification. In SQC, the glass product would be sampled after it is vitrified. Individual glass property-composition models form the basis for the “feed forward” SPC. The models transform constraints on the melt and glass properties into constraints on the feed composition going to the melter in order to guarantee, at the 95% confidence level, that the feed will be processable and that the durability of the resulting waste form will be acceptable to a geologic repository.

  4. Improving the process of process modelling by the use of domain process patterns

    NARCIS (Netherlands)

    Koschmider, A.; Reijers, H.A.

    2015-01-01

    The use of business process models has become prevalent in a wide area of enterprise applications. But while their popularity is expanding, concerns are growing with respect to their proper creation and maintenance. An obvious way to boost the efficiency of creating high-quality business process

  5. 2 TeV walking technirho at LHC?

    Directory of Open Access Journals (Sweden)

    Hidenori S. Fukano

    2015-11-01

    Full Text Available The ATLAS collaboration has recently reported an excess of about 2.5 σ global significance at around 2 TeV in the diboson channel with the boson-tagged fat dijets, which may imply a new resonance beyond the standard model. We provide a possible explanation of the excess as the isospin-triplet technivector mesons (technirhos, denoted as ρΠ±,3 of the walking technicolor in the case of the one-family model as a benchmark. As the effective theory for the walking technicolor at the scales relevant to the LHC experiment, we take a scale-invariant version of the hidden local symmetry model so constructed as to accommodate technipions, technivector mesons, and the technidilaton in such a way that the model respects spontaneously broken chiral and scale symmetries of the underlying walking technicolor. In particular, the technidilaton, a (pseudo Nambu–Goldstone boson of the (approximate scale symmetry predicted in the walking technicolor, has been shown to be successfully identified with the 125 GeV Higgs. Currently available LHC limits on those technihadrons are used to fix the couplings of technivector mesons to the standard-model fermions and weak gauge bosons. We find that the technirhos are mainly produced through the Drell–Yan process and predominantly decay to the dibosons, which accounts for the currently reported excess at around 2 TeV. The consistency with the electroweak precision test and other possible discovery channels of the 2 TeV technirhos are also addressed.

  6. First Experiences with Kinect v2 Sensor for Close Range 3d Modelling

    Science.gov (United States)

    Lachat, E.; Macher, H.; Mittet, M.-A.; Landes, T.; Grussenmeyer, P.

    2015-02-01

    RGB-D cameras, also known as range imaging cameras, are a recent generation of sensors. As they are suitable for measuring distances to objects at high frame rate, such sensors are increasingly used for 3D acquisitions, and more generally for applications in robotics or computer vision. This kind of sensors became popular especially since the Kinect v1 (Microsoft) arrived on the market in November 2010. In July 2014, Windows has released a new sensor, the Kinect for Windows v2 sensor, based on another technology as its first device. However, due to its initial development for video games, the quality assessment of this new device for 3D modelling represents a major investigation axis. In this paper first experiences with Kinect v2 sensor are related, and the ability of close range 3D modelling is investigated. For this purpose, error sources on output data as well as a calibration approach are presented.

  7. FIRST EXPERIENCES WITH KINECT V2 SENSOR FOR CLOSE RANGE 3D MODELLING

    Directory of Open Access Journals (Sweden)

    E. Lachat

    2015-02-01

    Full Text Available RGB-D cameras, also known as range imaging cameras, are a recent generation of sensors. As they are suitable for measuring distances to objects at high frame rate, such sensors are increasingly used for 3D acquisitions, and more generally for applications in robotics or computer vision. This kind of sensors became popular especially since the Kinect v1 (Microsoft arrived on the market in November 2010. In July 2014, Windows has released a new sensor, the Kinect for Windows v2 sensor, based on another technology as its first device. However, due to its initial development for video games, the quality assessment of this new device for 3D modelling represents a major investigation axis. In this paper first experiences with Kinect v2 sensor are related, and the ability of close range 3D modelling is investigated. For this purpose, error sources on output data as well as a calibration approach are presented.

  8. Research on Process-oriented Spatio-temporal Data Model

    Directory of Open Access Journals (Sweden)

    XUE Cunjin

    2016-02-01

    Full Text Available According to the analysis of the present status and existing problems of spatio-temporal data models developed in last 20 years,this paper proposes a process-oriented spatio-temporal data model (POSTDM,aiming at representing,organizing and storing continuity and gradual geographical entities. The dynamic geographical entities are graded and abstracted into process objects series from their intrinsic characteristics,which are process objects,process stage objects,process sequence objects and process state objects. The logical relationships among process entities are further studied and the structure of UML models and storage are also designed. In addition,through the mechanisms of continuity and gradual changes impliedly recorded by process objects,and the modes of their procedure interfaces offered by the customized ObjcetStorageTable,the POSTDM can carry out process representation,storage and dynamic analysis of continuity and gradual geographic entities. Taking a process organization and storage of marine data as an example,a prototype system (consisting of an object-relational database and a functional analysis platform is developed for validating and evaluating the model's practicability.

  9. Markov Decision Process Measurement Model.

    Science.gov (United States)

    LaMar, Michelle M

    2018-03-01

    Within-task actions can provide additional information on student competencies but are challenging to model. This paper explores the potential of using a cognitive model for decision making, the Markov decision process, to provide a mapping between within-task actions and latent traits of interest. Psychometric properties of the model are explored, and simulation studies report on parameter recovery within the context of a simple strategy game. The model is then applied to empirical data from an educational game. Estimates from the model are found to correlate more strongly with posttest results than a partial-credit IRT model based on outcome data alone.

  10. Stochastic process corrosion growth models for pipeline reliability

    International Nuclear Information System (INIS)

    Bazán, Felipe Alexander Vargas; Beck, André Teófilo

    2013-01-01

    Highlights: •Novel non-linear stochastic process corrosion growth model is proposed. •Corrosion rate modeled as random Poisson pulses. •Time to corrosion initiation and inherent time-variability properly represented. •Continuous corrosion growth histories obtained. •Model is shown to precisely fit actual corrosion data at two time points. -- Abstract: Linear random variable corrosion models are extensively employed in reliability analysis of pipelines. However, linear models grossly neglect well-known characteristics of the corrosion process. Herein, a non-linear model is proposed, where corrosion rate is represented as a Poisson square wave process. The resulting model represents inherent time-variability of corrosion growth, produces continuous growth and leads to mean growth at less-than-one power of time. Different corrosion models are adjusted to the same set of actual corrosion data for two inspections. The proposed non-linear random process corrosion growth model leads to the best fit to the data, while better representing problem physics

  11. Model feedstock supply processing plants

    Directory of Open Access Journals (Sweden)

    V. M. Bautin

    2013-01-01

    Full Text Available The model of raw providing the processing enterprises entering into vertically integrated structure on production and processing of dairy raw materials, differing by an orientation on achievement of cumulative effect by the integrated structure acting as criterion function which maximizing is reached by optimization of capacities, volumes of deliveries of raw materials and its qualitative characteristics, costs of industrial processing of raw materials and demand for dairy production is developed.

  12. Modelling of radial electric fields and currents during divertor plate biasing on TdeV

    International Nuclear Information System (INIS)

    Lachambre, J.L.; Quirion, B.; Boucher, C.

    1994-01-01

    A simple model based on non-ambipolar radial transport and planar sheath physics is used to describe the generation of radial electric fields and currents in the scrape-off layer of the Tokamak de Varennes (TdeV) during divertor plate biasing. In general, the calculated predictions compare favourably with TdeV results over a variety of plasma conditions and divertor magnetic configurations. Validated by the experiment, the model is used to study the scaling laws of perpendicular ion mobility and to test existing related theories. Finally, the model is proposed as a useful tool for the design and upgrade of biased divertors through optimization of the plate and throat geometry. (author). 35 refs, 16 figs, 1 tab

  13. Assessing healthcare process maturity: challenges of using a business process maturity model

    NARCIS (Netherlands)

    Tarhan, A.; Turetken, O.; van den Biggelaar, F.J.H.M.

    2015-01-01

    Doi: 10.4108/icst.pervasivehealth.2015.259105 The quality of healthcare services is influenced by the maturity of healthcare processes used to develop it. A maturity model is an instrument to assess and continually improve organizational processes. In the last decade, a number of maturity models

  14. Determination of Process Parameters for High-Density, Ti-6Al-4V Parts Using Additive Manufacturing

    Energy Technology Data Exchange (ETDEWEB)

    Kamath, C. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-08-10

    In our earlier work, we described an approach for determining the process parameters that re- sult in high-density parts manufactured using the additive-manufacturing process of selective laser melting (SLM). Our approach, which combines simple simulations and experiments, was demon- strated using 316L stainless steel. We have also used the approach successfully for several other materials. This short note summarizes the results of our work in determining process parameters for Ti-6Al-4V using a Concept Laser M2 system.

  15. Model-based reasoning and the control of process plants

    International Nuclear Information System (INIS)

    Vaelisuo, Heikki

    1993-02-01

    In addition to feedback control, safe and economic operation of industrial process plants requires discrete-event type logic control like for example automatic control sequences, interlocks, etc. A lot of complex routine reasoning is involved in the design and verification and validation (VandV) of such automatics. Similar reasoning tasks are encountered during plant operation in action planning and fault diagnosis. The low-level part of the required problem solving is so straightforward that it could be accomplished by a computer if only there were plant models which allow versatile mechanised reasoning. Such plant models and corresponding inference algorithms are the main subject of this report. Deep knowledge and qualitative modelling play an essential role in this work. Deep knowledge refers to mechanised reasoning based on the first principles of the phenomena in the problem domain. Qualitative modelling refers to knowledge representation formalism and related reasoning methods which allow solving problems on an abstraction level higher than for example traditional simulation and optimisation. Prolog is a commonly used platform for artificial intelligence (Al) applications. Constraint logic languages like CLP(R) and Prolog-III extend the scope of logic programming to numeric problem solving. In addition they allow a programming style which often reduces the computational complexity significantly. An approach to model-based reasoning implemented in constraint logic programming language CLP(R) is presented. The approach is based on some of the principles of QSIM, an algorithm for qualitative simulation. It is discussed how model-based reasoning can be applied in the design and VandV of plant automatics and in action planning during plant operation. A prototype tool called ISIR is discussed and some initial results obtained during the development of the tool are presented. The results presented originate from preliminary test results of the prototype obtained

  16. Computer modeling of lung cancer diagnosis-to-treatment process.

    Science.gov (United States)

    Ju, Feng; Lee, Hyo Kyung; Osarogiagbon, Raymond U; Yu, Xinhua; Faris, Nick; Li, Jingshan

    2015-08-01

    We introduce an example of a rigorous, quantitative method for quality improvement in lung cancer care-delivery. Computer process modeling methods are introduced for lung cancer diagnosis, staging and treatment selection process. Two types of process modeling techniques, discrete event simulation (DES) and analytical models, are briefly reviewed. Recent developments in DES are outlined and the necessary data and procedures to develop a DES model for lung cancer diagnosis, leading up to surgical treatment process are summarized. The analytical models include both Markov chain model and closed formulas. The Markov chain models with its application in healthcare are introduced and the approach to derive a lung cancer diagnosis process model is presented. Similarly, the procedure to derive closed formulas evaluating the diagnosis process performance is outlined. Finally, the pros and cons of these methods are discussed.

  17. Models for the a subunits of the Thermus thermophilus V/A-ATPase and Saccharomyces cerevisiae V-ATPase enzymes by cryo-EM and evolutionary covariance

    Science.gov (United States)

    Schep, Daniel G.; Rubinstein, John L.

    2016-01-01

    Rotary ATPases couple ATP synthesis or hydrolysis to proton translocation across a membrane. However, understanding proton translocation has been hampered by a lack of structural information for the membrane-embedded a subunit. The V/A-ATPase from the eubacterium Thermus thermophilus is similar in structure to the eukaryotic V-ATPase but has a simpler subunit composition and functions in vivo to synthesize ATP rather than pump protons. We determined the T. thermophilus V/A-ATPase structure by cryo-EM at 6.4 Å resolution. Evolutionary covariance analysis allowed tracing of the a subunit sequence within the map, providing a complete model of the rotary ATPase. Comparing the membrane-embedded regions of the T. thermophilus V/A-ATPase and eukaryotic V-ATPase from Saccharomyces cerevisiae allowed identification of the α-helices that belong to the a subunit and revealed the existence of previously unknown subunits in the eukaryotic enzyme. Subsequent evolutionary covariance analysis enabled construction of a model of the a subunit in the S. cerevisae V-ATPase that explains numerous biochemical studies of that enzyme. Comparing the two a subunit structures determined here with a structure of the distantly related a subunit from the bovine F-type ATP synthase revealed a conserved pattern of residues, suggesting a common mechanism for proton transport in all rotary ATPases. PMID:26951669

  18. Systems analysis programs for Hands-on integrated reliability evaluations (SAPHIRE) Version 5.0: Verification and validation (V ampersand V) manual. Volume 9

    International Nuclear Information System (INIS)

    Jones, J.L.; Calley, M.B.; Capps, E.L.; Zeigler, S.L.; Galyean, W.J.; Novack, S.D.; Smith, C.L.; Wolfram, L.M.

    1995-03-01

    A verification and validation (V ampersand V) process has been performed for the System Analysis Programs for Hands-on Integrated Reliability Evaluation (SAPHIRE) Version 5.0. SAPHIRE is a set of four computer programs that NRC developed for performing probabilistic risk assessments. They allow an analyst to perform many of the functions necessary to create, quantify, and evaluate the risk associated with a facility or process being analyzed. The programs are Integrated Reliability and Risk Analysis System (IRRAS) System Analysis and Risk Assessment (SARA), Models And Results Database (MAR-D), and Fault tree, Event tree, and Piping and instrumentation diagram (FEP) graphical editor. Intent of this program is to perform a V ampersand V of successive versions of SAPHIRE. Previous efforts have been the V ampersand V of SAPHIRE Version 4.0. The SAPHIRE 5.0 V ampersand V plan is based on the SAPHIRE 4.0 V ampersand V plan with revisions to incorporate lessons learned from the previous effort. Also, the SAPHIRE 5.0 vital and nonvital test procedures are based on the test procedures from SAPHIRE 4.0 with revisions to include the new SAPHIRE 5.0 features as well as to incorporate lessons learned from the previous effort. Most results from the testing were acceptable; however, some discrepancies between expected code operation and actual code operation were identified. Modifications made to SAPHIRE are identified

  19. Towards a universal competitive intelligence process model

    Directory of Open Access Journals (Sweden)

    Rene Pellissier

    2013-08-01

    Full Text Available Background: Competitive intelligence (CI provides actionable intelligence, which provides a competitive edge in enterprises. However, without proper process, it is difficult to develop actionable intelligence. There are disagreements about how the CI process should be structured. For CI professionals to focus on producing actionable intelligence, and to do so with simplicity, they need a common CI process model.Objectives: The purpose of this research is to review the current literature on CI, to look at the aims of identifying and analysing CI process models, and finally to propose a universal CI process model.Method: The study was qualitative in nature and content analysis was conducted on all identified sources establishing and analysing CI process models. To identify relevant literature, academic databases and search engines were used. Moreover, a review of references in related studies led to more relevant sources, the references of which were further reviewed and analysed. To ensure reliability, only peer-reviewed articles were used.Results: The findings reveal that the majority of scholars view the CI process as a cycle of interrelated phases. The output of one phase is the input of the next phase.Conclusion: The CI process is a cycle of interrelated phases. The output of one phase is the input of the next phase. These phases are influenced by the following factors: decision makers, process and structure, organisational awareness and culture, and feedback.

  20. Modeling and simulation of heterogeneous catalytic processes

    CERN Document Server

    Dixon, Anthony

    2014-01-01

    Heterogeneous catalysis and mathematical modeling are essential components of the continuing search for better utilization of raw materials and energy, with reduced impact on the environment. Numerical modeling of chemical systems has progressed rapidly due to increases in computer power, and is used extensively for analysis, design and development of catalytic reactors and processes. This book presents reviews of the state-of-the-art in modeling of heterogeneous catalytic reactors and processes. Reviews by leading authorities in the respective areas Up-to-date reviews of latest techniques in modeling of catalytic processes Mix of US and European authors, as well as academic/industrial/research institute perspectives Connections between computation and experimental methods in some of the chapters.

  1. Synergy of modeling processes in the area of soft and hard modeling

    Directory of Open Access Journals (Sweden)

    Sika Robert

    2017-01-01

    Full Text Available High complexity of production processes results in more frequent use of computer systems for their modeling and simulation. Process modeling helps to find optimal solution, verify some assumptions before implementation and eliminate errors. In practice, modeling of production processes concerns two areas: hard modeling (based on differential equations of mathematical physics and soft (based on existing data. In the paper the possibility of synergistic connection of these two approaches was indicated: it means hard modeling support based on the tools used in soft modeling. It aims at significant reducing the time in order to obtain final results with the use of hard modeling. Some test were carried out in the Calibrate module of NovaFlow&Solid (NF&S simulation system in the frame of thermal analysis (ATAS-cup. The authors tested output values forecasting in NF&S system (solidification time on the basis of variable parameters of the thermal model (heat conduction, specific heat, density. Collected data was used as an input to prepare soft model with the use of MLP (Multi-Layer Perceptron neural network regression model. The approach described above enable to reduce the time of production process modeling with use of hard modeling and should encourage production companies to use it.

  2. Self-exciting point process in modeling earthquake occurrences

    International Nuclear Information System (INIS)

    Pratiwi, H.; Slamet, I.; Respatiwulan; Saputro, D. R. S.

    2017-01-01

    In this paper, we present a procedure for modeling earthquake based on spatial-temporal point process. The magnitude distribution is expressed as truncated exponential and the event frequency is modeled with a spatial-temporal point process that is characterized uniquely by its associated conditional intensity process. The earthquakes can be regarded as point patterns that have a temporal clustering feature so we use self-exciting point process for modeling the conditional intensity function. The choice of main shocks is conducted via window algorithm by Gardner and Knopoff and the model can be fitted by maximum likelihood method for three random variables. (paper)

  3. Computer Modelling of Dynamic Processes

    Directory of Open Access Journals (Sweden)

    B. Rybakin

    2000-10-01

    Full Text Available Results of numerical modeling of dynamic problems are summed in the article up. These problems are characteristic for various areas of human activity, in particular for problem solving in ecology. The following problems are considered in the present work: computer modeling of dynamic effects on elastic-plastic bodies, calculation and determination of performances of gas streams in gas cleaning equipment, modeling of biogas formation processes.

  4. Modeling process-structure-property relationships for additive manufacturing

    Science.gov (United States)

    Yan, Wentao; Lin, Stephen; Kafka, Orion L.; Yu, Cheng; Liu, Zeliang; Lian, Yanping; Wolff, Sarah; Cao, Jian; Wagner, Gregory J.; Liu, Wing Kam

    2018-02-01

    This paper presents our latest work on comprehensive modeling of process-structure-property relationships for additive manufacturing (AM) materials, including using data-mining techniques to close the cycle of design-predict-optimize. To illustrate the processstructure relationship, the multi-scale multi-physics process modeling starts from the micro-scale to establish a mechanistic heat source model, to the meso-scale models of individual powder particle evolution, and finally to the macro-scale model to simulate the fabrication process of a complex product. To link structure and properties, a highefficiency mechanistic model, self-consistent clustering analyses, is developed to capture a variety of material response. The model incorporates factors such as voids, phase composition, inclusions, and grain structures, which are the differentiating features of AM metals. Furthermore, we propose data-mining as an effective solution for novel rapid design and optimization, which is motivated by the numerous influencing factors in the AM process. We believe this paper will provide a roadmap to advance AM fundamental understanding and guide the monitoring and advanced diagnostics of AM processing.

  5. Calibrating the sqHIMMELI v1.0 wetland methane emission model with hierarchical modeling and adaptive MCMC

    Science.gov (United States)

    Susiluoto, Jouni; Raivonen, Maarit; Backman, Leif; Laine, Marko; Makela, Jarmo; Peltola, Olli; Vesala, Timo; Aalto, Tuula

    2018-03-01

    Estimating methane (CH4) emissions from natural wetlands is complex, and the estimates contain large uncertainties. The models used for the task are typically heavily parameterized and the parameter values are not well known. In this study, we perform a Bayesian model calibration for a new wetland CH4 emission model to improve the quality of the predictions and to understand the limitations of such models.The detailed process model that we analyze contains descriptions for CH4 production from anaerobic respiration, CH4 oxidation, and gas transportation by diffusion, ebullition, and the aerenchyma cells of vascular plants. The processes are controlled by several tunable parameters. We use a hierarchical statistical model to describe the parameters and obtain the posterior distributions of the parameters and uncertainties in the processes with adaptive Markov chain Monte Carlo (MCMC), importance resampling, and time series analysis techniques. For the estimation, the analysis utilizes measurement data from the Siikaneva flux measurement site in southern Finland. The uncertainties related to the parameters and the modeled processes are described quantitatively. At the process level, the flux measurement data are able to constrain the CH4 production processes, methane oxidation, and the different gas transport processes. The posterior covariance structures explain how the parameters and the processes are related. Additionally, the flux and flux component uncertainties are analyzed both at the annual and daily levels. The parameter posterior densities obtained provide information regarding importance of the different processes, which is also useful for development of wetland methane emission models other than the square root HelsinkI Model of MEthane buiLd-up and emIssion for peatlands (sqHIMMELI). The hierarchical modeling allows us to assess the effects of some of the parameters on an annual basis. The results of the calibration and the cross validation suggest that

  6. Efficient querying of large process model repositories

    NARCIS (Netherlands)

    Jin, Tao; Wang, Jianmin; La Rosa, M.; Hofstede, ter A.H.M.; Wen, Lijie

    2013-01-01

    Recent years have seen an increased uptake of business process management technology in industries. This has resulted in organizations trying to manage large collections of business process models. One of the challenges facing these organizations concerns the retrieval of models from large business

  7. [GSH fermentation process modeling using entropy-criterion based RBF neural network model].

    Science.gov (United States)

    Tan, Zuoping; Wang, Shitong; Deng, Zhaohong; Du, Guocheng

    2008-05-01

    The prediction accuracy and generalization of GSH fermentation process modeling are often deteriorated by noise existing in the corresponding experimental data. In order to avoid this problem, we present a novel RBF neural network modeling approach based on entropy criterion. It considers the whole distribution structure of the training data set in the parameter learning process compared with the traditional MSE-criterion based parameter learning, and thus effectively avoids the weak generalization and over-learning. Then the proposed approach is applied to the GSH fermentation process modeling. Our results demonstrate that this proposed method has better prediction accuracy, generalization and robustness such that it offers a potential application merit for the GSH fermentation process modeling.

  8. Discovering Process Reference Models from Process Variants Using Clustering Techniques

    NARCIS (Netherlands)

    Li, C.; Reichert, M.U.; Wombacher, Andreas

    2008-01-01

    In today's dynamic business world, success of an enterprise increasingly depends on its ability to react to changes in a quick and flexible way. In response to this need, process-aware information systems (PAIS) emerged, which support the modeling, orchestration and monitoring of business processes

  9. A comparative study of inelastic scattering models at energy levels ranging from 0.5 keV to 10 keV

    Energy Technology Data Exchange (ETDEWEB)

    Hu, Chia-Yu [Department of Photonics, National Cheng Kung University, Tainan 701, Taiwan (China); Lin, Chun-Hung, E-mail: chlin@mail.ncku.edu.tw [Department of Photonics, National Cheng Kung University, Tainan 701, Taiwan (China); Advanced Optoelectronic Technology Center, National Cheng Kung University, Tainan 701, Taiwan (China)

    2017-03-01

    Six models, including a single-scattering model, four hybrid models, and one dielectric function model, were evaluated using Monte Carlo simulations for aluminum and copper at incident beam energies ranging from 0.5 keV to 10 keV. The inelastic mean free path, mean energy loss per unit path length, and backscattering coefficients obtained by these models are compared and discussed to understand the merits of the various models. ANOVA (analysis of variance) statistical models were used to quantify the effects of inelastic cross section and energy loss models on the basis of the simulated results deviation from the experimental data for the inelastic mean free path, the mean energy loss per unit path length, and the backscattering coefficient, as well as their correlations. This work in this study is believed to be the first application of ANOVA models towards evaluating inelastic electron beam scattering models. This approach is an improvement over the traditional approach which involves only visual estimation of the difference between the experimental data and simulated results. The data suggests that the optimization of the effective electron number per atom, binding energy, and cut-off energy of an inelastic model for different materials at different beam energies is more important than the selection of inelastic models for Monte Carlo electron scattering simulation. During the simulations, parameters in the equations should be tuned according to different materials for different beam energies rather than merely employing default parameters for an arbitrary material. Energy loss models and cross-section formulas are not the main factors influencing energy loss. Comparison of the deviation of the simulated results from the experimental data shows a significant correlation (p < 0.05) between the backscattering coefficient and energy loss per unit path length. The inclusion of backscattering electrons generated by both primary and secondary electrons for

  10. The Role(s) of Process Models in Design Practice

    DEFF Research Database (Denmark)

    Iversen, Søren; Jensen, Mads Kunø Nyegaard; Vistisen, Peter

    2018-01-01

    This paper investigates how design process models are implemented and used in design-driven organisations. The archetypical theoretical framing of process models, describe their primary role as guiding the design process, and assign roles and deliverables throughout the process. We hypothesise...... that the process models also take more communicative roles in practice, both in terms of creating an internal design rationale, as well as demystifying the black box of design thinking to external stakeholders. We investigate this hypothesis through an interview study of four major danish design......-driven organisations, and analyse the different roles their archetypical process models take in their organisations. The main contribution is the identification of three, often overlapping roles, which design process models showed to assume in design-driven organisations: process guidance, adding transparency...

  11. Causally nonseparable processes admitting a causal model

    International Nuclear Information System (INIS)

    Feix, Adrien; Araújo, Mateus; Brukner, Caslav

    2016-01-01

    A recent framework of quantum theory with no global causal order predicts the existence of ‘causally nonseparable’ processes. Some of these processes produce correlations incompatible with any causal order (they violate so-called ‘causal inequalities’ analogous to Bell inequalities ) while others do not (they admit a ‘causal model’ analogous to a local model ). Here we show for the first time that bipartite causally nonseparable processes with a causal model exist, and give evidence that they have no clear physical interpretation. We also provide an algorithm to generate processes of this kind and show that they have nonzero measure in the set of all processes. We demonstrate the existence of processes which stop violating causal inequalities but are still causally nonseparable when mixed with a certain amount of ‘white noise’. This is reminiscent of the behavior of Werner states in the context of entanglement and nonlocality. Finally, we provide numerical evidence for the existence of causally nonseparable processes which have a causal model even when extended with an entangled state shared among the parties. (paper)

  12. Input modelling of ASSERT-PV V2R8M1 for RUFIC fuel bundle

    Energy Technology Data Exchange (ETDEWEB)

    Park, Joo Hwan; Suk, Ho Chun

    2001-02-01

    This report describes the input modelling for subchannel analysis of CANFLEX-RU (RUFIC) fuel bundle which has been developed for an advanced fuel bundle of CANDU-6 reactor, using ASSERT-PV V2R8M1 code. Execution file of ASSERT-PV V2R8M1 code was recently transferred from AECL under JRDC agreement between KAERI and AECL. SSERT-PV V2R8M1 which is quite different from COBRA-IV-i code has been developed for thermalhydraulic analysis of CANDU-6 fuel channel by subchannel analysis method and updated so that 43-element CANDU fuel geometry can be applied. Hence, ASSERT code can be applied to the subchannel analysis of RUFIC fuel bundle. The present report was prepared for ASSERT input modelling of RUFIC fuel bundle. Since the ASSERT results highly depend on user's input modelling, the calculation results may be quite different among the user's input models. The objective of the present report is the preparation of detail description of the background information for input data and gives credibility of the calculation results.

  13. Input modelling of ASSERT-PV V2R8M1 for RUFIC fuel bundle

    International Nuclear Information System (INIS)

    Park, Joo Hwan; Suk, Ho Chun

    2001-02-01

    This report describes the input modelling for subchannel analysis of CANFLEX-RU (RUFIC) fuel bundle which has been developed for an advanced fuel bundle of CANDU-6 reactor, using ASSERT-PV V2R8M1 code. Execution file of ASSERT-PV V2R8M1 code was recently transferred from AECL under JRDC agreement between KAERI and AECL. SSERT-PV V2R8M1 which is quite different from COBRA-IV-i code has been developed for thermalhydraulic analysis of CANDU-6 fuel channel by subchannel analysis method and updated so that 43-element CANDU fuel geometry can be applied. Hence, ASSERT code can be applied to the subchannel analysis of RUFIC fuel bundle. The present report was prepared for ASSERT input modelling of RUFIC fuel bundle. Since the ASSERT results highly depend on user's input modelling, the calculation results may be quite different among the user's input models. The objective of the present report is the preparation of detail description of the background information for input data and gives credibility of the calculation results

  14. Towards simplification of hydrologic modeling: Identification of dominant processes

    Science.gov (United States)

    Markstrom, Steven; Hay, Lauren E.; Clark, Martyn P.

    2016-01-01

    The Precipitation–Runoff Modeling System (PRMS), a distributed-parameter hydrologic model, has been applied to the conterminous US (CONUS). Parameter sensitivity analysis was used to identify: (1) the sensitive input parameters and (2) particular model output variables that could be associated with the dominant hydrologic process(es). Sensitivity values of 35 PRMS calibration parameters were computed using the Fourier amplitude sensitivity test procedure on 110 000 independent hydrologically based spatial modeling units covering the CONUS and then summarized to process (snowmelt, surface runoff, infiltration, soil moisture, evapotranspiration, interflow, baseflow, and runoff) and model performance statistic (mean, coefficient of variation, and autoregressive lag 1). Identified parameters and processes provide insight into model performance at the location of each unit and allow the modeler to identify the most dominant process on the basis of which processes are associated with the most sensitive parameters. The results of this study indicate that: (1) the choice of performance statistic and output variables has a strong influence on parameter sensitivity, (2) the apparent model complexity to the modeler can be reduced by focusing on those processes that are associated with sensitive parameters and disregarding those that are not, (3) different processes require different numbers of parameters for simulation, and (4) some sensitive parameters influence only one hydrologic process, while others may influence many

  15. Neuroscientific Model of Motivational Process

    Directory of Open Access Journals (Sweden)

    Sung-Il eKim

    2013-03-01

    Full Text Available Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three subprocesses, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous subprocesses, namely reward-driven approach, value-based decision making, and goal-directed control. Reward-driven approach is the process in which motivation is generated by reward anticipation and selective approach behaviors toward reward. This process recruits the ventral striatum (reward area in which basic stimulus-action association is formed, and is classified as an automatic motivation to which relatively less attention is assigned. By contrast, value-based decision making is the process of evaluating various outcomes of actions, learning through positive prediction error, and calculating the value continuously. The striatum and the orbitofrontal cortex (valuation area play crucial roles in sustaining motivation. Lastly, the goal-directed control is the process of regulating motivation through cognitive control to achieve goals. This consciously controlled motivation is associated with higher-level cognitive functions such as planning, retaining the goal, monitoring the performance, and regulating action. The anterior cingulate cortex (attention area and the dorsolateral prefrontal cortex (cognitive control area are the main neural circuits related to regulation of motivation. These three subprocesses interact with each other by sending reward prediction error signals through dopaminergic pathway from the striatum and to the prefrontal cortex. The neuroscientific model of motivational process suggests several educational implications with regard to the generation, maintenance, and regulation of motivation to learn in the learning environment.

  16. Business Process Simulation: Requirements for Business and Resource Models

    Directory of Open Access Journals (Sweden)

    Audrius Rima

    2015-07-01

    Full Text Available The purpose of Business Process Model and Notation (BPMN is to provide easily understandable graphical representation of business process. Thus BPMN is widely used and applied in various areas one of them being a business process simulation. This paper addresses some BPMN model based business process simulation problems. The paper formulate requirements for business process and resource models in enabling their use for business process simulation.

  17. Model of diffusers / permeators for hydrogen processing

    International Nuclear Information System (INIS)

    Jacobs, W. D.; Hang, T.

    2008-01-01

    Palladium-silver (Pd-Ag) diffusers are mainstays of hydrogen processing. Diffusers separate hydrogen from inert species such as nitrogen, argon or helium. The tubing becomes permeable to hydrogen when heated to more than 250 C and a differential pressure is created across the membrane. The hydrogen diffuses better at higher temperatures. Experimental or experiential results have been the basis for determining or predicting a diffuser's performance. However, the process can be mathematically modeled, and comparison to experimental or other operating data can be utilized to improve the fit of the model. A reliable model-based diffuser system design is the goal which will have impacts on tritium and hydrogen processing. A computer model has been developed to solve the differential equations for diffusion given the operating boundary conditions. The model was compared to operating data for a low pressure diffuser system. The modeling approach and the results are presented in this paper. (authors)

  18. Search for light bosons in decays of the 125 GeV Higgs boson in proton-proton collisions at sqrt(s) = 8 TeV

    Energy Technology Data Exchange (ETDEWEB)

    Khachatryan, Vardan; et al.

    2017-01-08

    A search is presented for decays beyond the standard model of the 125 GeV Higgs bosons to a pair of light bosons, based on models with extended scalar sectors. Light boson masses between 5 and 62.5 GeV are probed in final states containing four tau leptons, two muons and two b quarks, or two muons and two tau leptons. The results are from data in proton-proton collisions corresponding to an integrated luminosity of 19.7 inverse-femtobarns, accumulated by the CMS experiment at the LHC at a center-of-mass energy of 8 TeV. No evidence for such exotic decays is found in the data. Upper limits are set on the product of the cross section and branching fraction for several signal processes. The results are also compared to predictions of two-Higgs-doublet models, including those with an additional scalar singlet.

  19. Figure-ground processing during fixational saccades in V1: indication for higher-order stability.

    Science.gov (United States)

    Gilad, Ariel; Pesoa, Yair; Ayzenshtat, Inbal; Slovin, Hamutal

    2014-02-26

    In a typical visual scene we continuously perceive a "figure" that is segregated from the surrounding "background" despite ongoing microsaccades and small saccades that are performed when attempting fixation (fixational saccades [FSs]). Previously reported neuronal correlates of figure-ground (FG) segregation in the primary visual cortex (V1) showed enhanced activity in the "figure" along with suppressed activity in the noisy "background." However, it is unknown how this FG modulation in V1 is affected by FSs. To investigate this question, we trained two monkeys to detect a contour embedded in a noisy background while simultaneously imaging V1 using voltage-sensitive dyes. During stimulus presentation, the monkeys typically performed 1-3 FSs, which displaced the contour over the retina. Using eye position and a 2D analytical model to map the stimulus onto V1, we were able to compute FG modulation before and after each FS. On the spatial cortical scale, we found that, after each FS, FG modulation follows the stimulus retinal displacement and "hops" within the V1 retinotopic map, suggesting visual instability. On the temporal scale, FG modulation is initiated in the new retinotopic position before it disappeared from the old retinotopic position. Moreover, the FG modulation developed faster after an FS, compared with after stimulus onset, which may contribute to visual stability of FG segregation, along the timeline of stimulus presentation. Therefore, despite spatial discontinuity of FG modulation in V1, the higher-order stability of FG modulation along time may enable our stable and continuous perception.

  20. Models of transport processes in concrete

    International Nuclear Information System (INIS)

    Pommersheim, J.M.; Clifton, J.R.

    1991-01-01

    An approach being considered by the US Nuclear Regulatory Commission for disposal of low-level radioactive waste is to place the waste forms in concrete vaults buried underground. The vaults would need a service life of 500 years. Approaches for predicting the service life of concrete of such vaults include the use of mathematical models. Mathematical models are presented in this report for the major degradation processes anticipated for the concrete vaults, which are corrosion of steel reinforcement, sulfate attack, acid attack, and leaching. The models mathematically represent rate controlling processes including diffusion, convection, and reaction and sorption of chemical species. These models can form the basis for predicting the life of concrete under in-service conditions. 33 refs., 6 figs., 7 tabs

  1. Toward Cognitively Constrained Models of Language Processing: A Review

    Directory of Open Access Journals (Sweden)

    Margreet Vogelzang

    2017-09-01

    Full Text Available Language processing is not an isolated capacity, but is embedded in other aspects of our cognition. However, it is still largely unexplored to what extent and how language processing interacts with general cognitive resources. This question can be investigated with cognitively constrained computational models, which simulate the cognitive processes involved in language processing. The theoretical claims implemented in cognitive models interact with general architectural constraints such as memory limitations. This way, it generates new predictions that can be tested in experiments, thus generating new data that can give rise to new theoretical insights. This theory-model-experiment cycle is a promising method for investigating aspects of language processing that are difficult to investigate with more traditional experimental techniques. This review specifically examines the language processing models of Lewis and Vasishth (2005, Reitter et al. (2011, and Van Rij et al. (2010, all implemented in the cognitive architecture Adaptive Control of Thought—Rational (Anderson et al., 2004. These models are all limited by the assumptions about cognitive capacities provided by the cognitive architecture, but use different linguistic approaches. Because of this, their comparison provides insight into the extent to which assumptions about general cognitive resources influence concretely implemented models of linguistic competence. For example, the sheer speed and accuracy of human language processing is a current challenge in the field of cognitive modeling, as it does not seem to adhere to the same memory and processing capacities that have been found in other cognitive processes. Architecture-based cognitive models of language processing may be able to make explicit which language-specific resources are needed to acquire and process natural language. The review sheds light on cognitively constrained models of language processing from two angles: we

  2. Beyond dual-process models: A categorisation of processes underlying intuitive judgement and decision making

    NARCIS (Netherlands)

    Glöckner, A.; Witteman, C.L.M.

    2010-01-01

    Intuitive-automatic processes are crucial for making judgements and decisions. The fascinating complexity of these processes has attracted many decision researchers, prompting them to start investigating intuition empirically and to develop numerous models. Dual-process models assume a clear

  3. Modelling of transport and biogeochemical processes in pollution plumes: Literature review of model development

    DEFF Research Database (Denmark)

    Brun, A.; Engesgaard, Peter Knudegaard

    2002-01-01

    A literature survey shows how biogeochemical (coupled organic and inorganic reaction processes) transport models are based on considering the complete biodegradation process as either a single- or as a two-step process. It is demonstrated that some two-step process models rely on the Partial...... Equilibrium Approach (PEA). The PEA assumes the organic degradation step, and not the electron acceptor consumption step, is rate limiting. This distinction is not possible in one-step process models, where consumption of both the electron donor and acceptor are treated kinetically. A three-dimensional, two......-step PEA model is developed. The model allows for Monod kinetics and biomass growth, features usually included only in one-step process models. The biogeochemical part of the model is tested for a batch system with degradation of organic matter under the consumption of a sequence of electron acceptors...

  4. TeV gravity at neutrino telescopes

    International Nuclear Information System (INIS)

    Illana, J.I.; Masip, M.; Meloni, D.

    2005-01-01

    Cosmogenic neutrinos reach the Earth with energies around 10 9 GeV, and their interactions with matter will be measured in upcoming experiments (Auger, IceCube). Models with extra dimensions and the fundamental scale at the TeV could imply signals in these experiments. In particular, the production of microscopic black holes by cosmogenic neutrinos has been extensively studied in the literature. Here we make a complete analysis of gravity-mediated interactions at larger distances, where they can be calculated in the eikonal approximation. In these processes a neutrino of energy E ν interacts elastically with a parton inside a nucleon, loses a small fraction y of its energy, and starts a hadronic shower of energy yE ν ν . We analyze the ultraviolet dependence and the relevance of graviton emission in these processes, and show that they are negligible. We also study the energy distribution of cosmogenic events in AMANDA and IceCube and the possibility of multiple-bang events. For any neutrino flux, the observation of an enhanced rate of neutral current events above 100 TeV in neutrino telescopes could be explained by TeV-gravity interactions. The values of the fundamental scale of gravity that IceCube could reach are comparable to those to be explored at the LHC

  5. Pulsar TeV Halos Explain the Diffuse TeV Excess Observed by Milagro.

    Science.gov (United States)

    Linden, Tim; Buckman, Benjamin J

    2018-03-23

    Milagro observations have found bright, diffuse TeV emission concentrated along the galactic plane of the Milky Way. The intensity and spectrum of this emission is difficult to explain with current models of hadronic γ-ray production, and has been named the "TeV excess." We show that TeV emission from pulsars naturally explains this excess. Recent observations have detected "TeV halos" surrounding pulsars that are either nearby or particularly luminous. Extrapolating this emission to the full population of Milky Way pulsars indicates that the ensemble of "subthreshold" sources necessarily produces bright TeV emission diffusively along the Milky Way plane. Models indicate that the TeV halo γ-ray flux exceeds that from hadronic γ rays above an energy of ∼500  GeV. Moreover, the spectrum and intensity of TeV halo emission naturally matches the TeV excess. Finally, we show that upcoming HAWC observations will resolve a significant fraction of the TeV excess into individual TeV halos, conclusively confirming, or ruling out, this model.

  6. Dual processing model of medical decision-making

    Science.gov (United States)

    2012-01-01

    Background Dual processing theory of human cognition postulates that reasoning and decision-making can be described as a function of both an intuitive, experiential, affective system (system I) and/or an analytical, deliberative (system II) processing system. To date no formal descriptive model of medical decision-making based on dual processing theory has been developed. Here we postulate such a model and apply it to a common clinical situation: whether treatment should be administered to the patient who may or may not have a disease. Methods We developed a mathematical model in which we linked a recently proposed descriptive psychological model of cognition with the threshold model of medical decision-making and show how this approach can be used to better understand decision-making at the bedside and explain the widespread variation in treatments observed in clinical practice. Results We show that physician’s beliefs about whether to treat at higher (lower) probability levels compared to the prescriptive therapeutic thresholds obtained via system II processing is moderated by system I and the ratio of benefit and harms as evaluated by both system I and II. Under some conditions, the system I decision maker’s threshold may dramatically drop below the expected utility threshold derived by system II. This can explain the overtreatment often seen in the contemporary practice. The opposite can also occur as in the situations where empirical evidence is considered unreliable, or when cognitive processes of decision-makers are biased through recent experience: the threshold will increase relative to the normative threshold value derived via system II using expected utility threshold. This inclination for the higher diagnostic certainty may, in turn, explain undertreatment that is also documented in the current medical practice. Conclusions We have developed the first dual processing model of medical decision-making that has potential to enrich the current medical

  7. Dual processing model of medical decision-making.

    Science.gov (United States)

    Djulbegovic, Benjamin; Hozo, Iztok; Beckstead, Jason; Tsalatsanis, Athanasios; Pauker, Stephen G

    2012-09-03

    Dual processing theory of human cognition postulates that reasoning and decision-making can be described as a function of both an intuitive, experiential, affective system (system I) and/or an analytical, deliberative (system II) processing system. To date no formal descriptive model of medical decision-making based on dual processing theory has been developed. Here we postulate such a model and apply it to a common clinical situation: whether treatment should be administered to the patient who may or may not have a disease. We developed a mathematical model in which we linked a recently proposed descriptive psychological model of cognition with the threshold model of medical decision-making and show how this approach can be used to better understand decision-making at the bedside and explain the widespread variation in treatments observed in clinical practice. We show that physician's beliefs about whether to treat at higher (lower) probability levels compared to the prescriptive therapeutic thresholds obtained via system II processing is moderated by system I and the ratio of benefit and harms as evaluated by both system I and II. Under some conditions, the system I decision maker's threshold may dramatically drop below the expected utility threshold derived by system II. This can explain the overtreatment often seen in the contemporary practice. The opposite can also occur as in the situations where empirical evidence is considered unreliable, or when cognitive processes of decision-makers are biased through recent experience: the threshold will increase relative to the normative threshold value derived via system II using expected utility threshold. This inclination for the higher diagnostic certainty may, in turn, explain undertreatment that is also documented in the current medical practice. We have developed the first dual processing model of medical decision-making that has potential to enrich the current medical decision-making field, which is still to the

  8. Modeling nonhomogeneous Markov processes via time transformation.

    Science.gov (United States)

    Hubbard, R A; Inoue, L Y T; Fann, J R

    2008-09-01

    Longitudinal studies are a powerful tool for characterizing the course of chronic disease. These studies are usually carried out with subjects observed at periodic visits giving rise to panel data. Under this observation scheme the exact times of disease state transitions and sequence of disease states visited are unknown and Markov process models are often used to describe disease progression. Most applications of Markov process models rely on the assumption of time homogeneity, that is, that the transition rates are constant over time. This assumption is not satisfied when transition rates depend on time from the process origin. However, limited statistical tools are available for dealing with nonhomogeneity. We propose models in which the time scale of a nonhomogeneous Markov process is transformed to an operational time scale on which the process is homogeneous. We develop a method for jointly estimating the time transformation and the transition intensity matrix for the time transformed homogeneous process. We assess maximum likelihood estimation using the Fisher scoring algorithm via simulation studies and compare performance of our method to homogeneous and piecewise homogeneous models. We apply our methodology to a study of delirium progression in a cohort of stem cell transplantation recipients and show that our method identifies temporal trends in delirium incidence and recovery.

  9. Mathematical modeling of biomass fuels formation process

    International Nuclear Information System (INIS)

    Gaska, Krzysztof; Wandrasz, Andrzej J.

    2008-01-01

    The increasing demand for thermal and electric energy in many branches of industry and municipal management accounts for a drastic diminishing of natural resources (fossil fuels). Meanwhile, in numerous technical processes, a huge mass of wastes is produced. A segregated and converted combustible fraction of the wastes, with relatively high calorific value, may be used as a component of formed fuels. The utilization of the formed fuel components from segregated groups of waste in associated processes of co-combustion with conventional fuels causes significant savings resulting from partial replacement of fossil fuels, and reduction of environmental pollution resulting directly from the limitation of waste migration to the environment (soil, atmospheric air, surface and underground water). The realization of technological processes with the utilization of formed fuel in associated thermal systems should be qualified by technical criteria, which means that elementary processes as well as factors of sustainable development, from a global viewpoint, must not be disturbed. The utilization of post-process waste should be preceded by detailed technical, ecological and economic analyses. In order to optimize the mixing process of fuel components, a mathematical model of the forming process was created. The model is defined as a group of data structures which uniquely identify a real process and conversion of this data in algorithms based on a problem of linear programming. The paper also presents the optimization of parameters in the process of forming fuels using a modified simplex algorithm with a polynomial worktime. This model is a datum-point in the numerical modeling of real processes, allowing a precise determination of the optimal elementary composition of formed fuels components, with assumed constraints and decision variables of the task

  10. Discrete processes modelling and geometry description in RTS and T code

    International Nuclear Information System (INIS)

    Degtyarev, I.I.; Liashenko, O.A.; Lokhovitskii, A.E.; Yazynin, I.A.; Belyakov-Bodin, V.I.; Blokhin, A.I.

    2001-01-01

    This paper describes the recent version of the RTS and T code system. RTS and T performs detailed simulations of many types of particles transport in complex 3D geometries in the energy range from a part of eV up to 20 TeV. A description of the main processes is given. (author)

  11. Process Model for Friction Stir Welding

    Science.gov (United States)

    Adams, Glynn

    1996-01-01

    Friction stir welding (FSW) is a relatively new process being applied for joining of metal alloys. The process was initially developed by The Welding Institute (TWI) in Cambridge, UK. The FSW process is being investigated at NASA/MSEC as a repair/initial weld procedure for fabrication of the super-light-weight aluminum-lithium shuttle external tank. The FSW investigations at MSFC were conducted on a horizontal mill to produce butt welds of flat plate material. The weldment plates are butted together and fixed to a backing plate on the mill bed. A pin tool is placed into the tool holder of the mill spindle and rotated at approximately 400 rpm. The pin tool is then plunged into the plates such that the center of the probe lies at, one end of the line of contact, between the plates and the shoulder of the pin tool penetrates the top surface of the weldment. The weld is produced by traversing the tool along the line of contact between the plates. A lead angle allows the leading edge of the shoulder to remain above the top surface of the plate. The work presented here is the first attempt at modeling a complex phenomenon. The mechanical aspects of conducting the weld process are easily defined and the process itself is controlled by relatively few input parameters. However, in the region of the weld, plasticizing and forging of the parent material occurs. These are difficult processes to model. The model presented here addresses only variations in the radial dimension outward from the pin tool axis. Examinations of the grain structure of the weld reveal that a considerable amount of material deformation also occurs in the direction parallel to the pin tool axis of rotation, through the material thickness. In addition, measurements of the axial load on the pin tool demonstrate that the forging affect of the pin tool shoulder is an important process phenomenon. Therefore, the model needs to be expanded to account for the deformations through the material thickness and the

  12. Statistical Model Analysis of (n, α Cross Sections for 4.0-6.5 MeV Neutrons

    Directory of Open Access Journals (Sweden)

    Khuukhenkhuu G.

    2016-01-01

    Full Text Available The statistical model based on the Weisskopf-Ewing theory and constant nuclear temperature approximation is used for systematical analysis of the 4.0-6.5 MeV neutron induced (n, α reaction cross sections. The α-clusterization effect was considered in the (n, α cross sections. A certain dependence of the (n, α cross sections on the relative neutron excess parameter of the target nuclei was observed. The systematic regularity of the (n, α cross sections behaviour is useful to estimate the same reaction cross sections for unstable isotopes. The results of our analysis can be used for nuclear astrophysical calculations such as helium burning and possible branching in the s-process.

  13. Measurement of 197Au(tau,xnyp) excitation functions for 15 MeV <= Esub(tau) <= 135 MeV and analysis in the framework of the hybrid model

    International Nuclear Information System (INIS)

    Bousshid, O.

    1981-01-01

    The (tau,xnyp)-reactions on 197 Au were measured. The 3 He incident energy was between 15 MeV and 135 MeV. The experiments were carried out using the stacked-foils technique. Cross sections were determind from the activity of the residual nuclei. The (tau,xn)-excitation functions were measured for 2 = 70 MeV as well as x >= 7 were measured for the first time. Further the (tau,pxn)- and (tau,2pxn)-excitation functions, which were not known so far, have now been measured. The analysis within the framework of the hybrid model for precompound-nuclear-reactions followed by an evaporation cascade, resulted in the best agreement between experimental data and theoretical model calculation using an initial exciton number nsub(o) = 5 (1n+3p+1h). The region of validity of the hybrid model for complex projectiles is discussed. (orig.) [de

  14. Modeling carbon sequestration in afforestation, agroforestry and forest management projects: the CO2FIX V.2 approach

    NARCIS (Netherlands)

    Masera, O.R.; Garza-Caligaris, J.F.; Kanninen, M.; Karjalainen, T.; Liski, J.; Nabuurs, G.J.; Pussinen, A.; Jong de, B.H.J.; Mohren, G.M.J.

    2003-01-01

    The paper describes the Version 2 of the CO2FIX (CO2FIX V.2) model, a user-friendly tool for dynamically estimating the carbon sequestration potential of forest management, agroforesty and afforestation projects. CO2FIX V.2 is a multi-cohort ecosystem-level model based on carbon accounting of forest

  15. Towards a structured process modeling method: Building the prescriptive modeling theory information on submission

    NARCIS (Netherlands)

    Claes, J.; Vanderfeesten, I.T.P.; Gailly, F.; Grefen, P.W.P.J.; Poels, G.

    2017-01-01

    In their effort to control and manage processes, organizations often create process models. The quality of such models is not always optimal, because it is challenging for a modeler to translate her mental image of the process into a formal process description. In order to support this complex human

  16. The Model of the Production Process for the Quality Management

    Directory of Open Access Journals (Sweden)

    Alot Zbigniew

    2017-02-01

    Full Text Available This article is a result of the research on the models of the production processes for the quality management and their identification. It discusses the classical model and the indicators for evaluating the capabilities by taking as its starting point the assumption of the normal distribution of the process characteristics. The division of the process types proposed by ISO 21747:2006 standard introducing models for non-stationary processes is presented. A general process model that allows in any real case to precisely describe the statistical characteristics of the process is proposed. It gives the opportunity for more detailed description, in comparison to the model proposed by ISO 21747:2006 standard, of the process characteristics and determining its capability. This model contains the type of process, statistical distribution, and the method for determining the capability and performance (long-term capability of the process. One of the model elements is proposed, own classification and resulting set of process types. The classification follows the recommendations of ISO 21747:2006 introducing models for the non-stationary processes. However, the set of the process types allows, beyond a more precise description of the process characteristics, its usage to monitor the process.

  17. Aspect-Oriented Business Process Modeling with AO4BPMN

    Science.gov (United States)

    Charfi, Anis; Müller, Heiko; Mezini, Mira

    Many crosscutting concerns in business processes need to be addressed already at the business process modeling level such as compliance, auditing, billing, and separation of duties. However, existing business process modeling languages including OMG's Business Process Modeling Notation (BPMN) lack appropriate means for expressing such concerns in a modular way. In this paper, we motivate the need for aspect-oriented concepts in business process modeling languages and propose an aspect-oriented extension to BPMN called AO4BPMN. We also present a graphical editor supporting that extension.

  18. Dual processing model of medical decision-making

    OpenAIRE

    Djulbegovic, Benjamin; Hozo, Iztok; Beckstead, Jason; Tsalatsanis, Athanasios; Pauker, Stephen G

    2012-01-01

    Abstract Background Dual processing theory of human cognition postulates that reasoning and decision-making can be described as a function of both an intuitive, experiential, affective system (system I) and/or an analytical, deliberative (system II) processing system. To date no formal descriptive model of medical decision-making based on dual processing theory has been developed. Here we postulate such a model and apply it to a common clinical situation: whether treatment should be administe...

  19. Stretchable V2O5/PEDOT supercapacitors: a modular fabrication process and charging with triboelectric nanogenerators.

    Science.gov (United States)

    Qi, Ruijie; Nie, Jinhui; Liu, Mingyang; Xia, Mengyang; Lu, Xianmao

    2018-04-26

    Stretchable energy storage devices are of great importance for the viable applications of wearable/stretchable electronics. Studies on stretchable energy storage devices, especially supercapacitors (SCs), have shown encouraging progress. However, challenges still remain in the pursuit of high specific capacitances and facile fabrication methods. Herein, we report a modular materials fabrication and assembly process for stretchable SCs. With a V2O5/PEDOT composite as the active material, the resulting stretchable SCs exhibited high areal specific capacitances up to 240 mF cm-2 and good capacitance retention at a strain of 50%. To demonstrate the facile assembly process, a stretchable wristband was fabricated by simply assembling SC cells in series to deliver a voltage higher than 2 V. Charging the wristband with a triboelectric nanogenerator (TENG) to light an LED was further demonstrated, indicating the potential to integrate our SCs with environmental energy harvesters for self-powered stretchable devices.

  20. Data-driven process decomposition and robust online distributed modelling for large-scale processes

    Science.gov (United States)

    Shu, Zhang; Lijuan, Li; Lijuan, Yao; Shipin, Yang; Tao, Zou

    2018-02-01

    With the increasing attention of networked control, system decomposition and distributed models show significant importance in the implementation of model-based control strategy. In this paper, a data-driven system decomposition and online distributed subsystem modelling algorithm was proposed for large-scale chemical processes. The key controlled variables are first partitioned by affinity propagation clustering algorithm into several clusters. Each cluster can be regarded as a subsystem. Then the inputs of each subsystem are selected by offline canonical correlation analysis between all process variables and its controlled variables. Process decomposition is then realised after the screening of input and output variables. When the system decomposition is finished, the online subsystem modelling can be carried out by recursively block-wise renewing the samples. The proposed algorithm was applied in the Tennessee Eastman process and the validity was verified.

  1. Application of mathematical and computer methods in mine rescue. Primeneniye matematicheskikh metodov i EVM v gornospasatel'nom dele

    Energy Technology Data Exchange (ETDEWEB)

    1980-01-01

    The collection contains the following articles: ''Control automation in mine rescue''. Khudosovtzev N.M., Teper V.B., Shschukin O.F.; ''Selection of optimal control of fire-fighting equipment''. Kozlyuk A.I., Charkov V.P., Teper V.B.; ''Formal approach to describing the process of underground fire extinguishing''. Kozlyu, A.I., Teper V.B., Romanchuk A.L.,; ''Computer selection of rescue and accident prevention methods using situation control methods''. Starovoitov V.T., Tomarovshschenko V.D., Koifman Yu. L., Plepler R.,; ''Planning experiments for studying the foam generation process using a steam-gas mixture''. Karyagina N.V., Makarenko V.L., Zhukova N.E., Shkodskikh V.I.; ''Mathematical model of the temperature field of spontaneous coal combustion in a mined out area''. Chuntu G.I., Kalyusskii A.E., Lysenko E.P., Gusar G.A.; ''Mathematical model of the cooling process of rock by a gas flow''. Kleiner A.A., Semko V.N.; ''M/G/I system with selfcontained servicing''. Skripnik V.M.; ''Closed-circuit servicing with mobile one-channel devices''. Skripnik V.M.; ''Analysis of methods for describing accident situations when modelling mine rescue operations''. Shschukin O.F., Brusentzev G.K., Laktionov O.A.; ''PROLOG system and the solution to several artificial intelligence problems''. Belov V.N., Branovitzkii V.I., Getzko L.N., Kudryavtzeva S.P., Seraya V.V.; ''Representation of knowledge in interactive systems''. Branovitzkii V.I., Getzko L.N., Kudryavtzeva S.P., Sakhno A.A., Seraya V.V.; ''Programming the selection of fire extinguishing equipment for rapid control of underground fire fighting using a computer''. Teper V.B., Ibchenko I.B., Bychkov G.D.; ''Interactive method to assess the effectiveness of exogenous fire fighting methods''. Brusentzev G.K., Shschukin O.F., Laktionov O.A., Shulga O.N.; ''Practical instruction using the automatic 'mine rescuer' system''.

  2. Ontological Analysis of Integrated Process Models: testing hypotheses

    Directory of Open Access Journals (Sweden)

    Michael Rosemann

    2001-11-01

    Full Text Available Integrated process modeling is achieving prominence in helping to document and manage business administration and IT processes in organizations. The ARIS framework is a popular example for a framework of integrated process modeling not least because it underlies the 800 or more reference models embedded in the world's most popular ERP package, SAP R/3. This paper demonstrates the usefulness of the Bunge-Wand-Weber (BWW representation model for evaluating modeling grammars such as those constituting ARIS. It reports some initial insights gained from pilot testing Green and Rosemann's (2000 evaluative propositions. Even when considering all five views of ARIS, modelers have problems representing business rules, the scope and boundary of systems, and decomposing models. However, even though it is completely ontologically redundant, users still find the function view useful in modeling.

  3. TeV scale leptogenesis in B−L model with alternative cosmologies

    Energy Technology Data Exchange (ETDEWEB)

    Abdallah, W., E-mail: wabdallah@zewailcity.edu.eg [Centre for Theoretical Physics, Zewail City of Science and Technology, Sheikh Zayed, 12588, Giza (Egypt); Department of Mathematics, Faculty of Science, Cairo University, Giza (Egypt); Delepine, D., E-mail: delepine@fisica.ugto.mx [Division de Ciencias e Ingenierías, Universidad de Guanajuato, C.P. 37150, León, Guanajuato (Mexico); Khalil, S., E-mail: skhalil@zewailcity.edu.eg [Centre for Theoretical Physics, Zewail City of Science and Technology, Sheikh Zayed, 12588, Giza (Egypt); Department of Mathematics, Faculty of Science, Ain Shams University, Cairo (Egypt); School of Physics and Astronomy, University of Southampton, Highfield, Southampton SO17 1BJ (United Kingdom)

    2013-10-01

    In TeV scale B−L extension of the standard model with inverse seesaw, the Yukawa coupling of right-handed neutrinos can be of order one. This implies that the out-of-equilibrium condition for leptogenesis within the standard cosmology is not satisfied. We provide two scenarios for overcoming this problem and generating the desired value of the baryon asymmetry of the Universe. The first scenario is based on the extra-dimensional braneworld effects that modify the Friedman equation. We show that in this case the value of the baryon asymmetry of the Universe constrains the five-dimensional Planck mass to be of order 100 TeV. In the second scenario a non-thermal right-handed neutrino produced by the decay of a heavy moduli is assumed. In this case, we emphasize that it is possible to generate the required baryon asymmetry of the Universe for TeV scale right-handed neutrinos.

  4. Computer-Aided Multiscale Modelling for Chemical Process Engineering

    DEFF Research Database (Denmark)

    Morales Rodriguez, Ricardo; Gani, Rafiqul

    2007-01-01

    Chemical processes are generally modeled through monoscale approaches, which, while not adequate, satisfy a useful role in product-process design. In this case, use of a multi-dimensional and multi-scale model-based approach has importance in product-process development. A computer-aided framework...

  5. Guided interaction exploration in artifact-centric process models

    NARCIS (Netherlands)

    van Eck, M.L.; Sidorova, N.; van der Aalst, W.M.P.

    2017-01-01

    Artifact-centric process models aim to describe complex processes as a collection of interacting artifacts. Recent development in process mining allow for the discovery of such models. However, the focus is often on the representation of the individual artifacts rather than their interactions. Based

  6. Modeling keV particle interactions with molecular and polymeric samples

    International Nuclear Information System (INIS)

    Delcorte, Arnaud

    2005-01-01

    Organic surfaces are locally submitted to extreme, out of equilibrium conditions when they are bombarded by kiloelectronvolt particles (atoms, ions, clusters). The time scale of the energy transfer is from tens of femtoseconds to several picoseconds depending on the material and the average energy per atom in the energized volume is of the order of a few eV, i.e. sufficient to break bonds in the solid. As a result, atoms, molecules and their fragments are released in the gas phase, which makes sputtering/desorption methods useful for surface treatment (ion beam patterning) and analysis (mass spectrometry). The radicals created in the sample also induce branching and cross-linking reactions that can be useful for surface modification purposes. Molecular dynamics simulations have provided an invaluable help for the elucidation of keV particle-induced processes in organic overlayers and, most recently, bulk materials. In this review, I illustrate the various mechanisms at play using case studies taken from our recent investigations and from the literature. They include the Ar-induced sputtering of a large polymeric molecule on a metal substrate and a molecular sample made of polystyrene oligomers. The emphasis is placed on the understanding of the energy transfer processes in the disturbed surface region and the mechanisms of molecule desorption, fragmentation and recombination, crucial for ion beam-based analytical methods

  7. Towards simplification of hydrologic modeling: identification of dominant processes

    Directory of Open Access Journals (Sweden)

    S. L. Markstrom

    2016-11-01

    Full Text Available parameter hydrologic model, has been applied to the conterminous US (CONUS. Parameter sensitivity analysis was used to identify: (1 the sensitive input parameters and (2 particular model output variables that could be associated with the dominant hydrologic process(es. Sensitivity values of 35 PRMS calibration parameters were computed using the Fourier amplitude sensitivity test procedure on 110 000 independent hydrologically based spatial modeling units covering the CONUS and then summarized to process (snowmelt, surface runoff, infiltration, soil moisture, evapotranspiration, interflow, baseflow, and runoff and model performance statistic (mean, coefficient of variation, and autoregressive lag 1. Identified parameters and processes provide insight into model performance at the location of each unit and allow the modeler to identify the most dominant process on the basis of which processes are associated with the most sensitive parameters. The results of this study indicate that: (1 the choice of performance statistic and output variables has a strong influence on parameter sensitivity, (2 the apparent model complexity to the modeler can be reduced by focusing on those processes that are associated with sensitive parameters and disregarding those that are not, (3 different processes require different numbers of parameters for simulation, and (4 some sensitive parameters influence only one hydrologic process, while others may influence many.

  8. BUSINESS PROCESS MODELLING: A FOUNDATION FOR KNOWLEDGE MANAGEMENT

    Directory of Open Access Journals (Sweden)

    Vesna Bosilj-Vukšić

    2006-12-01

    Full Text Available Knowledge management (KM is increasingly recognised as a strategic practice of knowledge-intensive companies, becoming an integral part of an organisation's strategy to improve business performance. This paper provides an overview of business process modelling applications and analyses the relationship between business process modelling and knowledge management projects. It presents the case study of Croatian leading banks and the insurance company, discussing its practical experience in conducting business process modelling projects and investigating the opportunity for integrating business process repository and organisational knowledge as the foundation for knowledge management system development.

  9. Towards a universal competitive intelligence process model

    Directory of Open Access Journals (Sweden)

    Rene Pellissier

    2013-08-01

    Objectives: The purpose of this research is to review the current literature on CI, to look at the aims of identifying and analysing CI process models, and finally to propose a universal CI process model. Method: The study was qualitative in nature and content analysis was conducted on all identified sources establishing and analysing CI process models. To identify relevant literature, academic databases and search engines were used. Moreover, a review of references in related studies led to more relevant sources, the references of which were further reviewed and analysed. To ensure reliability, only peer-reviewed articles were used. Results: The findings reveal that the majority of scholars view the CI process as a cycle of interrelated phases. The output of one phase is the input of the next phase. Conclusion: The CI process is a cycle of interrelated phases. The output of one phase is the input of the next phase. These phases are influenced by the following factors: decision makers, process and structure, organisational awareness and culture, and feedback.

  10. Simulation of engine auxiliary drive V-belt slip motion. Part 1. Development of belt slip model; Engine hoki V belt slip kyodo no simulation. 1. Belt slip model no kaihatsu

    Energy Technology Data Exchange (ETDEWEB)

    Kurisu, T [Mazda Motor Corp., Hiroshima (Japan)

    1997-10-01

    V-belts are widely used for driving auxiliary components of an engine. Inadequet design of such belt system sometimes results in troubles such as belt squeak, side rubber separation and/or bottom rubber crack. However, there has been no design tools which can predict belt slip quantitatively. The author developed a motion simulation program of Auxiliary Drive V-Belt System considering belt slip. The program showed good prediction accuracy for belt slip motion. This paper describes the simulation model. 1 ref., 12 figs.

  11. Pedagogic process modeling: Humanistic-integrative approach

    Directory of Open Access Journals (Sweden)

    Boritko Nikolaj M.

    2007-01-01

    Full Text Available The paper deals with some current problems of modeling the dynamics of the subject-features development of the individual. The term "process" is considered in the context of the humanistic-integrative approach, in which the principles of self education are regarded as criteria for efficient pedagogic activity. Four basic characteristics of the pedagogic process are pointed out: intentionality reflects logicality and regularity of the development of the process; discreteness (stageability in dicates qualitative stages through which the pedagogic phenomenon passes; nonlinearity explains the crisis character of pedagogic processes and reveals inner factors of self-development; situationality requires a selection of pedagogic conditions in accordance with the inner factors, which would enable steering the pedagogic process. Offered are two steps for singling out a particular stage and the algorithm for developing an integrative model for it. The suggested conclusions might be of use for further theoretic research, analyses of educational practices and for realistic predicting of pedagogical phenomena. .

  12. A protocol for the intercomparison of marine fishery and ecosystem models: Fish-MIP v1.0

    Science.gov (United States)

    Tittensor, Derek P.; Eddy, Tyler D.; Lotze, Heike K.; Galbraith, Eric D.; Cheung, William; Barange, Manuel; Blanchard, Julia L.; Bopp, Laurent; Bryndum-Buchholz, Andrea; Büchner, Matthias; Bulman, Catherine; Carozza, David A.; Christensen, Villy; Coll, Marta; Dunne, John P.; Fernandes, Jose A.; Fulton, Elizabeth A.; Hobday, Alistair J.; Huber, Veronika; Jennings, Simon; Jones, Miranda; Lehodey, Patrick; Link, Jason S.; Mackinson, Steve; Maury, Olivier; Niiranen, Susa; Oliveros-Ramos, Ricardo; Roy, Tilla; Schewe, Jacob; Shin, Yunne-Jai; Silva, Tiago; Stock, Charles A.; Steenbeek, Jeroen; Underwood, Philip J.; Volkholz, Jan; Watson, James R.; Walker, Nicola D.

    2018-04-01

    Model intercomparison studies in the climate and Earth sciences communities have been crucial to building credibility and coherence for future projections. They have quantified variability among models, spurred model development, contrasted within- and among-model uncertainty, assessed model fits to historical data, and provided ensemble projections of future change under specified scenarios. Given the speed and magnitude of anthropogenic change in the marine environment and the consequent effects on food security, biodiversity, marine industries, and society, the time is ripe for similar comparisons among models of fisheries and marine ecosystems. Here, we describe the Fisheries and Marine Ecosystem Model Intercomparison Project protocol version 1.0 (Fish-MIP v1.0), part of the Inter-Sectoral Impact Model Intercomparison Project (ISIMIP), which is a cross-sectoral network of climate impact modellers. Given the complexity of the marine ecosystem, this class of models has substantial heterogeneity of purpose, scope, theoretical underpinning, processes considered, parameterizations, resolution (grain size), and spatial extent. This heterogeneity reflects the lack of a unified understanding of the marine ecosystem and implies that the assemblage of all models is more likely to include a greater number of relevant processes than any single model. The current Fish-MIP protocol is designed to allow these heterogeneous models to be forced with common Earth System Model (ESM) Coupled Model Intercomparison Project Phase 5 (CMIP5) outputs under prescribed scenarios for historic (from the 1950s) and future (to 2100) time periods; it will be adapted to CMIP phase 6 (CMIP6) in future iterations. It also describes a standardized set of outputs for each participating Fish-MIP model to produce. This enables the broad characterization of differences between and uncertainties within models and projections when assessing climate and fisheries impacts on marine ecosystems and the

  13. A protocol for the intercomparison of marine fishery and ecosystem models: Fish-MIP v1.0

    Directory of Open Access Journals (Sweden)

    D. P. Tittensor

    2018-04-01

    Full Text Available Model intercomparison studies in the climate and Earth sciences communities have been crucial to building credibility and coherence for future projections. They have quantified variability among models, spurred model development, contrasted within- and among-model uncertainty, assessed model fits to historical data, and provided ensemble projections of future change under specified scenarios. Given the speed and magnitude of anthropogenic change in the marine environment and the consequent effects on food security, biodiversity, marine industries, and society, the time is ripe for similar comparisons among models of fisheries and marine ecosystems. Here, we describe the Fisheries and Marine Ecosystem Model Intercomparison Project protocol version 1.0 (Fish-MIP v1.0, part of the Inter-Sectoral Impact Model Intercomparison Project (ISIMIP, which is a cross-sectoral network of climate impact modellers. Given the complexity of the marine ecosystem, this class of models has substantial heterogeneity of purpose, scope, theoretical underpinning, processes considered, parameterizations, resolution (grain size, and spatial extent. This heterogeneity reflects the lack of a unified understanding of the marine ecosystem and implies that the assemblage of all models is more likely to include a greater number of relevant processes than any single model. The current Fish-MIP protocol is designed to allow these heterogeneous models to be forced with common Earth System Model (ESM Coupled Model Intercomparison Project Phase 5 (CMIP5 outputs under prescribed scenarios for historic (from the 1950s and future (to 2100 time periods; it will be adapted to CMIP phase 6 (CMIP6 in future iterations. It also describes a standardized set of outputs for each participating Fish-MIP model to produce. This enables the broad characterization of differences between and uncertainties within models and projections when assessing climate and fisheries impacts on marine ecosystems

  14. Updating parameters of the chicken processing line model

    DEFF Research Database (Denmark)

    Kurowicka, Dorota; Nauta, Maarten; Jozwiak, Katarzyna

    2010-01-01

    A mathematical model of chicken processing that quantitatively describes the transmission of Campylobacter on chicken carcasses from slaughter to chicken meat product has been developed in Nauta et al. (2005). This model was quantified with expert judgment. Recent availability of data allows...... updating parameters of the model to better describe processes observed in slaughterhouses. We propose Bayesian updating as a suitable technique to update expert judgment with microbiological data. Berrang and Dickens’s data are used to demonstrate performance of this method in updating parameters...... of the chicken processing line model....

  15. Kinetic and thermodynamic modelling of TBP synthesis processes

    International Nuclear Information System (INIS)

    Azzouz, A.; Attou, M.

    1989-02-01

    The present paper deals with kinetic and thermodynamic modellisation of tributylphosphate (TBP) synthesis processes. Its aim consists in a purely comparative study of two different synthesis ways i.e. direct and indirect estirification of butanol. The methodology involves two steps. The first step consists in approximating curves which describe the process evolution and their dependence on the main parameters. The results gave a kinetic model of the process rate yielding in TBP. Further, on the basis of thermodynamic data concerning the various involved compounds a theoretical model was achieved. The calculations were carried out in Basic language and an interpolation mathematical method was applied to approximate the kinetic curves. The thermodynamic calculations were achieved on the basis of GIBBS' free energy using a VAX type computer and a VT240 terminal. The calculations accuracy was reasonable and within the norms. For each process, the confrontation of both models leads to an appreciable accord. In the two processes, the thermodynamic models were similar although the kinetic equations present different reaction orders. Hence the reaction orders were determined by a mathematical method which conists in searching the minimal difference between an empiric relation and a kinetic model with fixed order. This corresponds in fact in testing the model proposed at various reaction order around the suspected value. The main idea which results from such a work is that this kind of processes is well fitting with the model without taking into account the side chain reactions. The process behaviour is like that of a single reaction having a quasi linear dependence of the rate yielding and the reaction time for both processes

  16. On the correlation between process model metrics and errors

    NARCIS (Netherlands)

    Mendling, J.; Neumann, G.; Aalst, van der W.M.P.; Grundy, J.; Hartmann, S.; Laender, S.; Maciaszek, L.; Roddick, J.F.

    2007-01-01

    Business process models play an important role for the management, design, and improvement of process organizations and process-aware information systems. Despite the extensive application of process modeling in practice there are hardly empirical results available on quality aspects of process

  17. Improved regional-scale groundwater representation by the coupling of the mesoscale Hydrologic Model (mHM v5.7) to the groundwater model OpenGeoSys (OGS)

    Science.gov (United States)

    Jing, Miao; Heße, Falk; Kumar, Rohini; Wang, Wenqing; Fischer, Thomas; Walther, Marc; Zink, Matthias; Zech, Alraune; Samaniego, Luis; Kolditz, Olaf; Attinger, Sabine

    2018-06-01

    Most large-scale hydrologic models fall short in reproducing groundwater head dynamics and simulating transport process due to their oversimplified representation of groundwater flow. In this study, we aim to extend the applicability of the mesoscale Hydrologic Model (mHM v5.7) to subsurface hydrology by coupling it with the porous media simulator OpenGeoSys (OGS). The two models are one-way coupled through model interfaces GIS2FEM and RIV2FEM, by which the grid-based fluxes of groundwater recharge and the river-groundwater exchange generated by mHM are converted to fixed-flux boundary conditions of the groundwater model OGS. Specifically, the grid-based vertical reservoirs in mHM are completely preserved for the estimation of land-surface fluxes, while OGS acts as a plug-in to the original mHM modeling framework for groundwater flow and transport modeling. The applicability of the coupled model (mHM-OGS v1.0) is evaluated by a case study in the central European mesoscale river basin - Nägelstedt. Different time steps, i.e., daily in mHM and monthly in OGS, are used to account for fast surface flow and slow groundwater flow. Model calibration is conducted following a two-step procedure using discharge for mHM and long-term mean of groundwater head measurements for OGS. Based on the model summary statistics, namely the Nash-Sutcliffe model efficiency (NSE), the mean absolute error (MAE), and the interquartile range error (QRE), the coupled model is able to satisfactorily represent the dynamics of discharge and groundwater heads at several locations across the study basin. Our exemplary calculations show that the one-way coupled model can take advantage of the spatially explicit modeling capabilities of surface and groundwater hydrologic models and provide an adequate representation of the spatiotemporal behaviors of groundwater storage and heads, thus making it a valuable tool for addressing water resources and management problems.

  18. Comparison of Different Turbulence Models for Numerical Simulation of Pressure Distribution in V-Shaped Stepped Spillway

    Directory of Open Access Journals (Sweden)

    Zhaoliang Bai

    2017-01-01

    Full Text Available V-shaped stepped spillway is a new shaped stepped spillway, and the pressure distribution is quite different from that of the traditional stepped spillway. In this paper, five turbulence models were used to simulate the pressure distribution in the skimming flow regimes. Through comparing with the physical value, the realizable k-ε model had better precision in simulating the pressure distribution. Then, the flow pattern of V-shaped and traditional stepped spillways was given to illustrate the unique pressure distribution using realizable k-ε turbulence model.

  19. Transparent predictive modelling of the twin screw granulation process using a compensated interval type-2 fuzzy system.

    Science.gov (United States)

    AlAlaween, Wafa' H; Khorsheed, Bilal; Mahfouf, Mahdi; Gabbott, Ian; Reynolds, Gavin K; Salman, Agba D

    2018-03-01

    In this research, a new systematic modelling framework which uses machine learning for describing the granulation process is presented. First, an interval type-2 fuzzy model is elicited in order to predict the properties of the granules produced by twin screw granulation (TSG) in the pharmaceutical industry. Second, a Gaussian mixture model (GMM) is integrated in the framework in order to characterize the error residuals emanating from the fuzzy model. This is done to refine the model by taking into account uncertainties and/or any other unmodelled behaviour, stochastic or otherwise. All proposed modelling algorithms were validated via a series of Laboratory-scale experiments. The size of the granules produced by TSG was successfully predicted, where most of the predictions fit within a 95% confidence interval. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. A generalized logarithmic image processing model based on the gigavision sensor model.

    Science.gov (United States)

    Deng, Guang

    2012-03-01

    The logarithmic image processing (LIP) model is a mathematical theory providing generalized linear operations for image processing. The gigavision sensor (GVS) is a new imaging device that can be described by a statistical model. In this paper, by studying these two seemingly unrelated models, we develop a generalized LIP (GLIP) model. With the LIP model being its special case, the GLIP model not only provides new insights into the LIP model but also defines new image representations and operations for solving general image processing problems that are not necessarily related to the GVS. A new parametric LIP model is also developed. To illustrate the application of the new scalar multiplication operation, we propose an energy-preserving algorithm for tone mapping, which is a necessary step in image dehazing. By comparing with results using two state-of-the-art algorithms, we show that the new scalar multiplication operation is an effective tool for tone mapping.