WorldWideScience

Sample records for application models based

  1. Modelling Gesture Based Ubiquitous Applications

    CERN Document Server

    Zacharia, Kurien; Varghese, Surekha Mariam

    2011-01-01

    A cost effective, gesture based modelling technique called Virtual Interactive Prototyping (VIP) is described in this paper. Prototyping is implemented by projecting a virtual model of the equipment to be prototyped. Users can interact with the virtual model like the original working equipment. For capturing and tracking the user interactions with the model image and sound processing techniques are used. VIP is a flexible and interactive prototyping method that has much application in ubiquitous computing environments. Different commercial as well as socio-economic applications and extension to interactive advertising of VIP are also discussed.

  2. Application software development via model based design

    OpenAIRE

    Haapala, O. (Olli)

    2015-01-01

    This thesis was set to study the utilization of the MathWorks’ Simulink® program in model based application software development and its compatibility with the Vacon 100 inverter. The target was to identify all the problems related to everyday usage of this method and create a white paper of how to execute a model based design to create a Vacon 100 compatible system software. Before this thesis was started, there was very little knowledge of the compatibility of this method. However durin...

  3. Model-based vision for space applications

    Science.gov (United States)

    Chaconas, Karen; Nashman, Marilyn; Lumia, Ronald

    1992-01-01

    This paper describes a method for tracking moving image features by combining spatial and temporal edge information with model based feature information. The algorithm updates the two-dimensional position of object features by correlating predicted model features with current image data. The results of the correlation process are used to compute an updated model. The algorithm makes use of a high temporal sampling rate with respect to spatial changes of the image features and operates in a real-time multiprocessing environment. Preliminary results demonstrate successful tracking for image feature velocities between 1.1 and 4.5 pixels every image frame. This work has applications for docking, assembly, retrieval of floating objects and a host of other space-related tasks.

  4. Python-Based Applications for Hydrogeological Modeling

    Science.gov (United States)

    Khambhammettu, P.

    2013-12-01

    Python is a general-purpose, high-level programming language whose design philosophy emphasizes code readability. Add-on packages supporting fast array computation (numpy), plotting (matplotlib), scientific /mathematical Functions (scipy), have resulted in a powerful ecosystem for scientists interested in exploratory data analysis, high-performance computing and data visualization. Three examples are provided to demonstrate the applicability of the Python environment in hydrogeological applications. Python programs were used to model an aquifer test and estimate aquifer parameters at a Superfund site. The aquifer test conducted at a Groundwater Circulation Well was modeled with the Python/FORTRAN-based TTIM Analytic Element Code. The aquifer parameters were estimated with PEST such that a good match was produced between the simulated and observed drawdowns. Python scripts were written to interface with PEST and visualize the results. A convolution-based approach was used to estimate source concentration histories based on observed concentrations at receptor locations. Unit Response Functions (URFs) that relate the receptor concentrations to a unit release at the source were derived with the ATRANS code. The impact of any releases at the source could then be estimated by convolving the source release history with the URFs. Python scripts were written to compute and visualize receptor concentrations for user-specified source histories. The framework provided a simple and elegant way to test various hypotheses about the site. A Python/FORTRAN-based program TYPECURVEGRID-Py was developed to compute and visualize groundwater elevations and drawdown through time in response to a regional uniform hydraulic gradient and the influence of pumping wells using either the Theis solution for a fully-confined aquifer or the Hantush-Jacob solution for a leaky confined aquifer. The program supports an arbitrary number of wells that can operate according to arbitrary schedules. The

  5. Model-based clustering using copulas with applications

    OpenAIRE

    Kosmidis, Ioannis; Karlis, Dimitris

    2014-01-01

    The majority of model-based clustering techniques is based on multivariate Normal models and their variants. In this paper copulas are used for the construction of flexible families of models for clustering applications. The use of copulas in model-based clustering offers two direct advantages over current methods: i) the appropriate choice of copulas provides the ability to obtain a range of exotic shapes for the clusters, and ii) the explicit choice of marginal distributions for the cluster...

  6. Measurement-based load modeling: Theory and application

    Institute of Scientific and Technical Information of China (English)

    MA; Jin; HAN; Dong; HE; RenMu

    2007-01-01

    Load model is one of the most important elements in power system operation and control. However, owing to its complexity, load modeling is still an open and very difficult problem. Summarizing our work on measurement-based load modeling in China for more than twenty years, this paper systematically introduces the mathematical theory and applications regarding the load modeling. The flow chart and algorithms for measurement-based load modeling are presented. A composite load model structure with 13 parameters is also proposed. Analysis results based on the trajectory sensitivity theory indicate the importance of the load model parameters for the identification. Case studies show the accuracy of the presented measurement-based load model. The load model thus built has been validated by field measurements all over China. Future working directions on measurement- based load modeling are also discussed in the paper.

  7. Model Checking-Based Testing of Web Applications

    Institute of Scientific and Technical Information of China (English)

    ZENG Hongwei; MIAO Huaikou

    2007-01-01

    A formal model representing the navigation behavior of a Web application as the Kripke structure is proposed and an approach that applies model checking to test case generation is presented. The Object Relation Diagram as the object model is employed to describe the object structure of a Web application design and can be translated into the behavior model. A key problem of model checking-based test generation for a Web application is how to construct a set of trap properties that intend to cause the violations of model checking against the behavior model and output of counterexamples used to construct the test sequences.We give an algorithm that derives trap properties from the object model with respect to node and edge coverage criteria.

  8. Formal model based methodology for developing software for nuclear applications

    International Nuclear Information System (INIS)

    The approach used in model based design is to build the model of the system in graphical/textual language. In older model based design approach, the correctness of the model is usually established by simulation. Simulation which is analogous to testing, cannot guarantee that the design meets the system requirements under all possible scenarios. This is however possible if the modeling language is based on formal semantics so that the developed model can be subjected to formal verification of properties based on specification. The verified model can then be translated into an implementation through reliable/verified code generator thereby reducing the necessity of low level testing. Such a methodology is admissible as per guidelines of IEC60880 standard applicable to software used in computer based systems performing category A functions in nuclear power plant and would also be acceptable for category B functions. In this article, the experience in implementation and formal verification of important controllers used in the process control system of a nuclear reactor. We have used The SCADE (Safety Critical System Analysis and Design Environment) environment to model the controllers. The modeling language used in SCADE is based on the synchronous dataflow model of computation. A set of safety properties has been verified using formal verification technique

  9. WWW Business Applications Based on the Cellular Model

    Institute of Scientific and Technical Information of China (English)

    Toshio Kodama; Tosiyasu L. Kunii; Yoichi Seki

    2008-01-01

    A cellular model based on the Incrementally Modular Abstraction Hierarchy (IMAH) is a novel model that can represent the architecture of and changes in cyberworlds, preserving invariants from a general level to a specific one. We have developed a data processing system called the Cellular Data System (CDS). In the development of business applications, you can prevent combinatorial explosion in the process of business design and testing by using CDS. In this paper, we have first designed and implemented wide-use algebra on the presentation level. Next, we have developed and verified the effectiveness of two general business applications using CDS: 1) a customer information management system, and 2) an estimate system.

  10. Data Warehouse Model For Mobile-Based Applications

    Directory of Open Access Journals (Sweden)

    Muhammad Shahbani Abu Bakar

    2016-06-01

    Full Text Available Analysis and design are very important roles in the Data Warehouse (DW system development and forms as a backbone of any successful or failure of the DW project. The emerging trends of analytic-based application required the DW system to be implemented in the mobile environment. However, current analysis and design approaches are based on existing DW environments that focusing on the deployment of the DW system in traditional web-based applications. This will create the limitations on user accessed and the used of analytical information by the decision makers. Consequently, this will prolong the adoption of analytic-based applications to the users and organizations. This research aims to suggest an approach for modeling the DW and design the DW system on the mobile environments. A variant dimension of modeling techniques was used to enhance the DW schemas in order to accommodate the requirements of mobile characteristics in the DW design. A proposed mobile DW system was evaluated by expert review, and support the success of mobile DW-based application implementation

  11. Intelligent control based on intelligent characteristic model and its application

    Institute of Scientific and Technical Information of China (English)

    吴宏鑫; 王迎春; 邢琰

    2003-01-01

    This paper presents a new intelligent control method based on intelligent characteristic model for a kind of complicated plant with nonlinearities and uncertainties, whose controlled output variables cannot be measured on line continuously. The basic idea of this method is to utilize intelligent techniques to form the characteristic model of the controlled plant according to the principle of combining the char-acteristics of the plant with the control requirements, and then to present a new design method of intelli-gent controller based on this characteristic model. First, the modeling principles and expression of the intelligent characteristic model are presented. Then based on description of the intelligent characteristic model, the design principles and methods of the intelligent controller composed of several open-loops and closed-loops sub controllers with qualitative and quantitative information are given. Finally, the ap-plication of this method in alumina concentration control in the real aluminum electrolytic process is in-troduced. It is proved in practice that the above methods not only are easy to implement in engineering design but also avoid the trial-and-error of general intelligent controllers. It has taken better effect in the following application: achieving long-term stable control of low alumina concentration and increasing the controlled ratio of anode effect greatly from 60% to 80%.

  12. Medical applications of model-based dynamic thermography

    Science.gov (United States)

    Nowakowski, Antoni; Kaczmarek, Mariusz; Ruminski, Jacek; Hryciuk, Marcin; Renkielska, Alicja; Grudzinski, Jacek; Siebert, Janusz; Jagielak, Dariusz; Rogowski, Jan; Roszak, Krzysztof; Stojek, Wojciech

    2001-03-01

    The proposal to use active thermography in medical diagnostics is promising in some applications concerning investigation of directly accessible parts of the human body. The combination of dynamic thermograms with thermal models of investigated structures gives attractive possibility to make internal structure reconstruction basing on different thermal properties of biological tissues. Measurements of temperature distribution synchronized with external light excitation allow registration of dynamic changes of local temperature dependent on heat exchange conditions. Preliminary results of active thermography applications in medicine are discussed. For skin and under- skin tissues an equivalent thermal model may be determined. For the assumed model its effective parameters may be reconstructed basing on the results of transient thermal processes. For known thermal diffusivity and conductivity of specific tissues the local thickness of a two or three layer structure may be calculated. Results of some medical cases as well as reference data of in vivo study on animals are presented. The method was also applied to evaluate the state of the human heart during the open chest cardio-surgical interventions. Reference studies of evoked heart infarct in pigs are referred, too. We see the proposed new in medical applications technique as a promising diagnostic tool. It is a fully non-invasive, clean, handy, fast and affordable method giving not only qualitative view of investigated surfaces but also an objective quantitative measurement result, accurate enough for many applications including fast screening of affected tissues.

  13. Application of Z-Number Based Modeling in Psychological Research

    Directory of Open Access Journals (Sweden)

    Rafik Aliev

    2015-01-01

    Full Text Available Pilates exercises have been shown beneficial impact on physical, physiological, and mental characteristics of human beings. In this paper, Z-number based fuzzy approach is applied for modeling the effect of Pilates exercises on motivation, attention, anxiety, and educational achievement. The measuring of psychological parameters is performed using internationally recognized instruments: Academic Motivation Scale (AMS, Test of Attention (D2 Test, and Spielberger’s Anxiety Test completed by students. The GPA of students was used as the measure of educational achievement. Application of Z-information modeling allows us to increase precision and reliability of data processing results in the presence of uncertainty of input data created from completed questionnaires. The basic steps of Z-number based modeling with numerical solutions are presented.

  14. Application of Z-Number Based Modeling in Psychological Research.

    Science.gov (United States)

    Aliev, Rafik; Memmedova, Konul

    2015-01-01

    Pilates exercises have been shown beneficial impact on physical, physiological, and mental characteristics of human beings. In this paper, Z-number based fuzzy approach is applied for modeling the effect of Pilates exercises on motivation, attention, anxiety, and educational achievement. The measuring of psychological parameters is performed using internationally recognized instruments: Academic Motivation Scale (AMS), Test of Attention (D2 Test), and Spielberger's Anxiety Test completed by students. The GPA of students was used as the measure of educational achievement. Application of Z-information modeling allows us to increase precision and reliability of data processing results in the presence of uncertainty of input data created from completed questionnaires. The basic steps of Z-number based modeling with numerical solutions are presented. PMID:26339231

  15. Voxel-based model and its application in advanced manufacturing

    Science.gov (United States)

    Wu, Xiaojun; Liu, Weijun; Wang, Tianran

    2004-03-01

    Traditionally, 3D models, even so called solid ones, can only represent the object's surface information, and the interior is regarded as homogeneous. In most applications, it is necessary to represent the interior structures and attributes of an object, such as materials, density and color, etc. Surface model is incapable of bearing this task. In this case, voxel model is a good choice. Voxelization is the process of converting a geometrically represented 3D object into a three dimensional volume of dataset. In this paper, an algorithm is proposed to voxelize the polygonal meshes ported from current CAD modeling packages into volume datasets based on the easily indexing property of Octree structure. The minimal distance to the feature voxel (or voxels) is taken as criterion to distribute different material compositions to get a new kind of material called FGM (functionally graded material), which is suitable for the interface of RPM (Rapid Prototyping Manufacturing).

  16. Photographic-based target models for LADAR applications

    Science.gov (United States)

    Jack, James T.; Delashmit, Walter H.

    2009-05-01

    A long standing need for the application of laser radar (LADAR) to a wider range of targets is a technique for creating a "target model" from target photographs. This is feasible since LADAR images are 3D and photographs at selected azimuth/elevation angles will allow the required models to be created. Preferred photographic images of a wide range of selected targets were specified and collected. These photographs were processed using code developed in house and some commercial software packages. These "models" were used in model-based automatic target recognition (ATR) algorithms. The ATR performance was excellent. This technique differs significantly from other techniques for creating target models. Those techniques require CAD models which are much harder to manipulate and contain extraneous detail. The technique in this paper develops the photographic-based target models in component form so that any component (e.g., turret of a tank) can be independently manipulated, such as rotating the turret. This new technique also allows models to be generated for targets for which no actual LADAR data has ever been collected. A summary of the steps used in the modeling process is as follows: start with a set of input photographs, calibrate the imagery into a 3D world space to generate points corresponding to target features, create target geometry by connecting points with surfaces, mark all co-located points in each image view and verify alignment of points, place in a 3D space, create models by creating surfaces (i.e., connect points with planar curves) and scale target into real-world coordinates.

  17. Real-time application of the drag based model

    Science.gov (United States)

    Žic, Tomislav; Temmer, Manuela; Vršnak, Bojan

    2016-04-01

    The drag-based model (DBM) is an analytical model which is usually used for calculating kinematics of coronal mass ejections (CMEs) in the interplanetary space, prediction of the CME arrival times and impact speeds at arbitrary targets in the heliosphere. The main assumption of the model is that beyond a distance of about 20 solar radii from the Sun, the drag is dominant in the interplanetary space. The previous version of DBM relied on the rough assumption of averaged, unperturbed and constant environmental conditions as well as constant CME properties throughout the entire interplanetary CME propagation. The continuation of our work consists of enhancing the model into a form which uses a time dependent and perturbed environment without constraints on CME properties and distance forecasting. The extension provides the possibility of application in various scenarios, such as automatic least-square fitting on initial CME kinematic data suitable for a real-time forecasting of CME kinematics, or embedding the DBM into pre-calculated interplanetary ambient conditions provided by advanced numerical simulations (for example, codes of ENLIL, EUHFORIA, etc.). A demonstration of the enhanced DBM is available on the web-site: http://www.geof.unizg.hr/~tzic/dbm.html. We acknowledge the support of European Social Fund under the "PoKRet" project.

  18. A Component-based Programming Model for Composite, Distributed Applications

    Science.gov (United States)

    Eidson, Thomas M.; Bushnell, Dennis M. (Technical Monitor)

    2001-01-01

    The nature of scientific programming is evolving to larger, composite applications that are composed of smaller element applications. These composite applications are more frequently being targeted for distributed, heterogeneous networks of computers. They are most likely programmed by a group of developers. Software component technology and computational frameworks are being proposed and developed to meet the programming requirements of these new applications. Historically, programming systems have had a hard time being accepted by the scientific programming community. In this paper, a programming model is outlined that attempts to organize the software component concepts and fundamental programming entities into programming abstractions that will be better understood by the application developers. The programming model is designed to support computational frameworks that manage many of the tedious programming details, but also that allow sufficient programmer control to design an accurate, high-performance application.

  19. Extending EMMS-based models to CFB boiler applications

    Institute of Scientific and Technical Information of China (English)

    Bona Lu; Nan Zhang; Wei Wang; Jinghai Li

    2012-01-01

    Recently,EMMS-based models are being widely applied in simulations of high-throughput circulating fluidized beds (CFBs) with fine particles.Its use for low flux systems,such as CFB boiler (CFBB),still remains unexplored.In this work,it has been found that the original definition of cluster diameter in EMMS model is unsuitable for simulations of the CFB boiler with low solids flux.To remedy this,we propose a new model of cluster diameter.The EMMS-based drag model (EMMS/matrix model) with this revised cluster definition is validated through the computational fluid dynamics (CFD) simulation of a CFB boiler.

  20. CAD-model-based vision for space applications

    Science.gov (United States)

    Shapiro, Linda G.

    1988-01-01

    A pose acquisition system operating in space must be able to perform well in a variety of different applications including automated guidance and inspections tasks with many different, but known objects. Since the space station is being designed with automation in mind, there will be CAD models of all the objects, including the station itself. The construction of vision models and procedures directly from the CAD models is the goal of this project. The system that is being designed and implementing must convert CAD models to vision models, predict visible features from a given view point from the vision models, construct view classes representing views of the objects, and use the view class model thus derived to rapidly determine the pose of the object from single images and/or stereo pairs.

  1. Application of model-based and knowledge-based measuring methods as analytical redundancy

    International Nuclear Information System (INIS)

    The safe operation of nuclear power plants requires the application of modern and intelligent methods of signal processing for the normal operation as well as for the management of accident conditions. Such modern and intelligent methods are model-based and knowledge-based ones being founded on analytical knowledge (mathematical models) as well as experiences (fuzzy information). In addition to the existing hardware redundancies analytical redundancies will be established with the help of these modern methods. These analytical redundancies support the operating staff during the decision-making. The design of a hybrid model-based and knowledge-based measuring method will be demonstrated by the example of a fuzzy-supported observer. Within the fuzzy-supported observer a classical linear observer is connected with a fuzzy-supported adaptation of the model matrices of the observer model. This application is realized for the estimation of the non-measurable variables as steam content and mixture level within pressure vessels with water-steam mixture during accidental depressurizations. For this example the existing non-linearities will be classified and the verification of the model will be explained. The advantages of the hybrid method in comparison to the classical model-based measuring methods will be demonstrated by the results of estimation. The consideration of the parameters which have an important influence on the non-linearities requires the inclusion of high-dimensional structures of fuzzy logic within the model-based measuring methods. Therefore methods will be presented which allow the conversion of these high-dimensional structures to two-dimensional structures of fuzzy logic. As an efficient solution of this problem a method based on cascaded fuzzy controllers will be presented. (author). 2 refs, 12 figs, 5 tabs

  2. Numerical modeling in electroporation-based biomedical applications

    OpenAIRE

    Pavšelj, Nataša; Miklavčič, Damijan

    2015-01-01

    Background. Numerous experiments have to be performed before a biomedical application is put to practical use in clinical environment. As a complementary work to in vitro, in vivo and medical experiments, we can use analytical and numerical models to represent, as realistically as possible, real biological phenomena of, in our case, electroporation. In this way we canevaluate different electrical parameters in advance, such as pulse amplitude, duration, number of pulses, or different electrod...

  3. Numerical modeling in electroporation-based biomedical applications:

    OpenAIRE

    Miklavčič, Damijan; Pavšelj, Nataša

    2008-01-01

    Background. Numerous experiments have to be performed before a biomedical application is put to practical use in clinical environment. As a complementary work to in vitro, in vivo and medical experiments, we can use analytical and numerical models to represent, as realistically as possible, real biological phenomena of, in our case, electroporation. In this way we canevaluate different electrical parameters in advance, such as pulse amplitude, duration, number of pulses, or different electrod...

  4. UML based modeling of medical applications workflow in maxillofacial surgery

    OpenAIRE

    Toma, M; Busam, A; Ortmaier, T; Raczkowsky, J.; Höpner, C; Marmulla1, R.

    2007-01-01

    This paper presents our research in medical workflow modeling for computer- and robot-based surgical intervention in maxillofacial surgery. Our goal is to provide a method for clinical workflow modeling including workflow definition for pre- and intra-operative steps, analysis of new methods for combining conventional surgical procedures with robot- and computer-assisted procedures and facilitate an easy implementation of hard- and software systems.

  5. Physical based Schottky barrier diode modeling for THz applications

    DEFF Research Database (Denmark)

    Yan, Lei; Krozer, Viktor; Michaelsen, Rasmus Schandorph; Djurhuus, Torsten; Johansen, Tom Keinicke

    2013-01-01

    In this work, a physical Schottky barrier diode model is presented. The model is based on physical parameters such as anode area, Ohmic contact area, doping profile from epitaxial (EPI) and substrate (SUB) layers, layer thicknesses, barrier height, specific contact resistance, and device...... temperature. The effects of barrier height lowering, nonlinear resistance from the EPI layer, and hot electron noise are all included for accurate characterization of the Schottky diode. To verify the diode model, measured I-V and C-V characteristics are compared with the simulation results. Due to the lack...

  6. Model Based Fault Detection in a Centrifugal Pump Application

    DEFF Research Database (Denmark)

    Kallesøe, Carsten; Cocquempot, Vincent; Izadi-Zamanabadi, Roozbeh

    2006-01-01

    A model based approach for fault detection in a centrifugal pump, driven by an induction motor, is proposed in this paper. The fault detection algorithm is derived using a combination of structural analysis, observer design and Analytical Redundancy Relation (ARR) design. Structural considerations...... the algorithm is capable of detecting four different faults in the mechanical and hydraulic parts of the pump....

  7. GIS application on spatial landslide analysis using statistical based models

    Science.gov (United States)

    Pradhan, Biswajeet; Lee, Saro; Buchroithner, Manfred F.

    2009-09-01

    This paper presents the assessment results of spatially based probabilistic three models using Geoinformation Techniques (GIT) for landslide susceptibility analysis at Penang Island in Malaysia. Landslide locations within the study areas were identified by interpreting aerial photographs, satellite images and supported with field surveys. Maps of the topography, soil type, lineaments and land cover were constructed from the spatial data sets. There are ten landslide related factors were extracted from the spatial database and the frequency ratio, fuzzy logic, and bivariate logistic regression coefficients of each factor was computed. Finally, landslide susceptibility maps were drawn for study area using frequency ratios, fuzzy logic and bivariate logistic regression models. For verification, the results of the analyses were compared with actual landslide locations in study area. The verification results show that bivariate logistic regression model provides slightly higher prediction accuracy than the frequency ratio and fuzzy logic models.

  8. Application of Z-Number Based Modeling in Psychological Research

    OpenAIRE

    Rafik Aliev; Konul Memmedova

    2015-01-01

    Pilates exercises have been shown beneficial impact on physical, physiological, and mental characteristics of human beings. In this paper, Z-number based fuzzy approach is applied for modeling the effect of Pilates exercises on motivation, attention, anxiety, and educational achievement. The measuring of psychological parameters is performed using internationally recognized instruments: Academic Motivation Scale (AMS), Test of Attention (D2 Test), and Spielberger’s Anxiety Test completed by s...

  9. A Systematic Review of Agent-Based Modelling and Simulation Applications in the Higher Education Domain

    Science.gov (United States)

    Gu, X.; Blackmore, K. L.

    2015-01-01

    This paper presents the results of a systematic review of agent-based modelling and simulation (ABMS) applications in the higher education (HE) domain. Agent-based modelling is a "bottom-up" modelling paradigm in which system-level behaviour (macro) is modelled through the behaviour of individual local-level agent interactions (micro).…

  10. A Model of Cloud Based Application Environment for Software Testing

    CERN Document Server

    Vengattaraman, T; Baskaran, R

    2010-01-01

    Cloud computing is an emerging platform of service computing designed for swift and dynamic delivery of assured computing resources. Cloud computing provide Service-Level Agreements (SLAs) for guaranteed uptime availability for enabling convenient and on-demand network access to the distributed and shared computing resources. Though the cloud computing paradigm holds its potential status in the field of distributed computing, cloud platforms are not yet to the attention of majority of the researchers and practitioners. More specifically, still the researchers and practitioners community has fragmented and imperfect knowledge on cloud computing principles and techniques. In this context, one of the primary motivations of the work presented in this paper is to reveal the versatile merits of cloud computing paradigm and hence the objective of this work is defined to bring out the remarkable significances of cloud computing paradigm through an application environment. In this work, a cloud computing model for sof...

  11. Middleware Based Model of Heterogeneous Systems for SCADA Distributed Applications

    Directory of Open Access Journals (Sweden)

    UNGUREAN, I.

    2010-05-01

    Full Text Available Infrastructure underlying the distributed information systems is heterogeneous and very complex. Middleware allows the development of distributed information systems, without knowing the functioning details of an infrastructure, by its abstracting. An essential issue on designing such systems is represented by choosing the middleware technologies. An architectural model of a SCADA system based on middleware is proposed in this paper. This system is formed of servers that centralize data and clients, which receive information from a server, thus allowing the chart displaying of such information. All these components own a specific functionality and can exchange information, by means of a middleware bus. A middleware bus signifies a software bus, where more middleware technologies can coexist.

  12. Erosion risk assessment in the southern Amazon - Data Preprocessing, data base application and process based modelling

    Science.gov (United States)

    Schindewolf, Marcus; Herrmann, Marie-Kristin; Herrmann, Anne-Katrin; Schultze, Nico; Amorim, Ricardo S. S.; Schmidt, Jürgen

    2015-04-01

    The study region along the BR 16 highway belongs to the "Deforestation Arc" at the southern border of the Amazon rainforest. At the same time, it incorporates a land use gradient as colonization started in the 1975-1990 in Central Mato Grosso in 1990 in northern Mato Grosso and most recently in 2004-2005 in southern Pará. Based on present knowledge soil erosion is one of the key driver of soil degradation. Hence, there is a strong need to implement soil erosion control measures in eroding landscapes. Planning and dimensioning of such measures require reliable and detailed information on the temporal and spatial distribution of soil loss, sediment transport and deposition. Soil erosion models are increasingly used, in order to simulate the physical processes involved and to predict the effects of soil erosion control measures. The process based EROSION 3D simulation model is used for surveying soil erosion and deposition on regional catchments. Although EROSION 3D is a widespread, extensively validated model, the application of the model on regional scale remains challenging due to the enormous data requirements and complex data processing operations. In this context the study includes the compilation, validation and generalisation of existing land use and soil data in order to generate a consistent EROSION 3D input datasets. As a part of this process a GIS-linked data base application allows to transfer the original soil and land use data into model specific parameter files. This combined methodology provides different risk assessment maps for certain demands on regional scale. Besides soil loss and sediment transport, sediment pass over points into surface water bodies and particle enrichment can be simulated using the EROSION 3D model. Thus the estimation of particle bound nutrient and pollutant inputs into surface water bodies becomes possible. The study ended up in a user-friendly, timesaving and improved software package for the simulation of soil loss and

  13. Applications Of Algebraic Image Operators To Model-Based Vision

    Science.gov (United States)

    Lerner, Bao-Ting; Morelli, Michael V.; Thomas, Hans J.

    1989-03-01

    This paper extends our previous research on a highly structured and compact algebraic representation of grey-level images. Addition and multiplication are defined for the set of all grey-level images, which can then be described as polynomials of two variables. Utilizing this new algebraic structure, we have devised an innovative, efficient edge detection scheme.We have developed a robust method for linear feature extraction by combining the techniques of a Hough transform and a line follower with this new edge detection scheme. The major advantage of this feature extractor is its general, object-independent nature. Target attributes, such as line segment lengths, intersections, angles of intersection, and endpoints are derived by the feature extraction algorithm and employed during model matching. The feature extractor and model matcher are being incorporated into a distributed robot control system. Model matching is accomplished using both top-down and bottom-up processing: a priori sensor and world model information are used to constrain the search of the image space for features, while extracted image information is used to update the model.

  14. NOVEL COMPONENT-BASED DEVELOPMENT MODEL FOR SIP-BASED MOBILE APPLICATION

    Directory of Open Access Journals (Sweden)

    Ahmed Barnawi

    2012-02-01

    Full Text Available Universities and Institutions these days’ deals with issues related to with assessment of large number ofstudents. Various evaluation methods have been adopted by examiners in different institutions to examiningthe ability of an individual, starting from manual means of using paper and pencil to electronic, from oralto written, practical to theoretical and many others.There is a need to expedite the process of examination in order to meet the increasing enrolment of studentsat the universities and institutes. Sip Based Mass Mobile Examination System (SiBMMES expedites theexamination process by automating various activities in an examination such as exam paper setting,Scheduling and allocating examination time and evaluation (auto-grading for objective questions etc.SiBMMES uses the IP Multimedia Subsystem (IMS that is an IP communications framework providing anenvironment for the rapid development of innovative and reusable services Session Initial Protocol (SIP isa signalling (request-response protocol for this architecture and it is used for establishing sessions in anIP network, making it an ideal candidate for supporting terminal mobility in the IMS to deliver the services,with the extended services available in IMS like open APIs, common network services, Quality of Services(QoS like multiple sessions per call, Push to Talk etc often requiring multiple types of media (includingvoice, video, pictures, and text. SiBMMES is an effective solution for mass education evaluation usingmobile and web technology.In this paper, a novel hybrid component based development (CBD model is proposed for SiBMMES. AComponent based Hybrid Model is selected to the fact that IMS takes the concept of layered architectureone step further by defining a horizontal architecture where service enablers and common functions can bereused for multiple applications. This novel model tackle a new domain for IT professionals, its ideal tostart developing services as a small

  15. Resilience-based application of state-and-transition models

    Science.gov (United States)

    We recommend that several conceptual modifications be incorporated into the state-and-transition model (STM) framework to: 1) explicitly link this framework to the concept of ecological resilience, 2) direct management attention away from thresholds and toward the maintenance of state resilience, an...

  16. Modeling Component-based Bragg gratings Application: tunable lasers

    Directory of Open Access Journals (Sweden)

    Hedara Rachida

    2011-09-01

    Full Text Available The principal function of a grating Bragg is filtering, which can be used in optical fibers based component and active or passive semi conductors based component, as well as telecommunication systems. Their ideal use is with lasers with fiber, amplifiers with fiber or Laser diodes. In this work, we are going to show the principal results obtained during the analysis of various types of grating Bragg by the method of the coupled modes. We then present the operation of DBR are tunable. The use of Bragg gratings in a laser provides single-mode sources, agile wavelength. The use of sampled grating increases the tuning range.

  17. The timing of disability insurance application: a choice-based semiparametric hazard model

    OpenAIRE

    Richard V. Burkhauser; Butler, J. S.; Yang-Woo Kim

    1996-01-01

    We use a choice-based subsample of Social Security Disability Insurance applicants from the 1978 Social Security Survey of Disability and Work to test the importance of policy variables on the timing of application for disability insurance benefits following the onset of a work limiting health condition. We correct for choice-based sampling by extending the Manski-Lerman (1977) correction to the likelihood function of our continuous time hazard model defined with semiparametric unmeasured het...

  18. Application for managing model-based material properties for simulation-based engineering

    Science.gov (United States)

    Hoffman, Edward L.

    2009-03-03

    An application for generating a property set associated with a constitutive model of a material includes a first program module adapted to receive test data associated with the material and to extract loading conditions from the test data. A material model driver is adapted to receive the loading conditions and a property set and operable in response to the loading conditions and the property set to generate a model response for the material. A numerical optimization module is adapted to receive the test data and the model response and operable in response to the test data and the model response to generate the property set.

  19. Applicability of an exposure model for the determination of emissions from mobile phone base stations

    DEFF Research Database (Denmark)

    Breckenkamp, J; Neitzke, H P; Bornkessel, C;

    2008-01-01

    Applicability of a model to estimate radiofrequency electromagnetic field (RF-EMF) strength in households from mobile phone base stations was evaluated with technical data of mobile phone base stations available from the German Net Agency, and dosimetric measurements, performed in an...

  20. The Model and Design for COM-Based e-Commerce Application System

    Institute of Scientific and Technical Information of China (English)

    TANG Xiao-mei; SUN Li

    2002-01-01

    From the point of constructing e-commerce application system, based on the structured analysis and ObjectOriented Design method, a combined modeling method Business-Process Driven(BPD) is proposed. This method focuses on the business process through the development process of the system. First, the business model of the system, then commercial object model is introduced according to the business model. At last the COM-model for the system is established. The system is implemented in an iterative and incremental way. The design and analysis result of each stage is illustrated by series of views using the modeling tool UML.

  1. A Comparative Study of Relational and Non-Relational Database Models in a Web- Based Application

    OpenAIRE

    Cornelia Gyorödi; Robert Gyorödi; Roxana Sotoc

    2015-01-01

    The purpose of this paper is to present a comparative study between relational and non-relational database models in a web-based application, by executing various operations on both relational and on non-relational databases thus highlighting the results obtained during performance comparison tests. The study was based on the implementation of a web-based application for population records. For the non-relational database, we used MongoDB and for the relational database, we used MSSQL 2014. W...

  2. Security Model for Microsoft Based Mobile Sales Management Application in Private Cloud Computing

    Directory of Open Access Journals (Sweden)

    Kuan Chee Houng

    2013-05-01

    Full Text Available The Microsoft-based mobile sales management application is a sales force management application that currently running on Windows Mobile 6.5. It handles sales-related activity and cuts down the administrative task of sales representative. Then, Windows launch a new mobile operating system, Windows Phone and stop providing support to Windows Mobile. This has become an obstacle for Windows Mobile development. From time to time, Windows Mobile will be eliminated from the market due to no support provided by Microsoft. Besides that, Windows Mobile application cannot run on Windows Phone mobile operating system due to lack of compatibility. Therefore, applications those run on Windows Mobile need to find a solution addressing this problem. The rise of cloud computing technology in delivering software as a service becomes a solution. The Microsoft-based mobile sales management application delivers a service to run in a web browser, rather than limited by certain type of mobile that run the Windows Mobile operating system. However, there are some security issues need to concern in order to deliver the Microsoft-based mobile application as a service in private cloud computing. Therefore, security model is needed to answer the security issues in private cloud computing. This research is to propose a security model for the Microsoft-based mobile sales management application in private cloud computing. Lastly, a User Acceptance Test (UAT is carried out to test the compatibility between proposed security model of Microsoft-based mobile sales management application in a private cloud and tablet computers.

  3. A QUESTIONNAIRE APPROACH BASED ON THE TECHNOLOGY ACCEPTANCE MODEL FOR MOBILE TRACKING ON PATIENT PROGRESS APPLICATIONS

    Directory of Open Access Journals (Sweden)

    Hussain Mohammad Abu-Dalbouh

    2013-01-01

    Full Text Available Healthcare professionals spend much of their time wandering between patients and offices, while the supportive technology stays stationary. Therefore, mobile applications has adapted for healthcare industry. In spite of the advancement and variety of available mobile based applications, there is an eminent need to investigate the current position of the acceptance of those mobile health applications that are tailored towards the tracking patients condition, share patients information and access. Consequently, in this study Technology Acceptance Model has designed to investigate the user acceptance of mobile technology application within healthcare industry. The purpose of this study is to design a quantitative approach based on the technology acceptance model questionnaire as its primary research methodology. It utilized a quantitative approach based a Technology Acceptance Model (TAM to evaluate the system mobile tracking Model. The related constructs for evaluation are: Perceived of Usefulness, Perceived Ease of Use, User Satisfaction and Attribute of Usability. All these constructs are modified to suit the context of the study. Moreover, this study outlines the details of each construct and its relevance toward the research issue. The outcome of the study represents series of approaches that will apply for checking the suitability of a mobile tracking on patient progress application for health care industry and how well it achieves the aims and objectives of the design.

  4. Application of soft computing based hybrid models in hydrological variables modeling: a comprehensive review

    Science.gov (United States)

    Fahimi, Farzad; Yaseen, Zaher Mundher; El-shafie, Ahmed

    2016-02-01

    Since the middle of the twentieth century, artificial intelligence (AI) models have been used widely in engineering and science problems. Water resource variable modeling and prediction are the most challenging issues in water engineering. Artificial neural network (ANN) is a common approach used to tackle this problem by using viable and efficient models. Numerous ANN models have been successfully developed to achieve more accurate results. In the current review, different ANN models in water resource applications and hydrological variable predictions are reviewed and outlined. In addition, recent hybrid models and their structures, input preprocessing, and optimization techniques are discussed and the results are compared with similar previous studies. Moreover, to achieve a comprehensive view of the literature, many articles that applied ANN models together with other techniques are included. Consequently, coupling procedure, model evaluation, and performance comparison of hybrid models with conventional ANN models are assessed, as well as, taxonomy and hybrid ANN models structures. Finally, current challenges and recommendations for future researches are indicated and new hybrid approaches are proposed.

  5. Model-Based Instruction: Theory and Application in a Teacher Education Program.

    Science.gov (United States)

    Steinley, Gary; Reisetter, Marcy; Penrod, Kathryn; Haar, Jean; Ellingson, Janna

    Model-Based Instruction (MBI) plays a significant role in the undergraduate teacher education program at South Dakota State University. Integrated into the program 8 years ago, the understandings and applications of MBI have evolved into a powerful and comprehensive framework that leads to rich and varied instruction with students directly in the…

  6. Application of Set Pair Analysis-Based Similarity Forecast Model and Wavelet Denoising for Runoff Forecasting

    OpenAIRE

    Chien-Ming Chou

    2014-01-01

    This study presents the application of a set pair analysis-based similarity forecast (SPA-SF) model and wavelet denoising to forecast annual runoff. The SPA-SF model was built from identical, discrepant and contrary viewpoints. The similarity between estimated and historical data can be obtained. The weighted average of the annual runoff values characterized by the highest connection coefficients was regarded as the predicted value of the estimated annual runoff. In addition, runoff time seri...

  7. CACM: A New Coordination Model in Mobile Agent-Based Information Retrieval Applications

    Institute of Scientific and Technical Information of China (English)

    TANGXinhuai; ZHANGYaying; YAOYinxiong; YOUJinyuan

    2005-01-01

    In mobile agent systems, an application may be composed of several mobile agents that cooperatively perform a task. Multiple mobile agents need to communicate and interact with each other to accomplish their cooperative goal. Coordination model aims to provide solutions to interactions between concurrent activities, hiding the computing details and focusing on interaction between activities. A Context-aware coordination model (CACM), which combines mobility and coordination, is proposed for mobile agent applications, i.e. in mobile agent based information retrieval applications. The context-aware coordination model transfers interactions between agents from globally coupling interactions to locally uncoupling tuple space interactions. In addition, programmable tuple space is adopted to solve the problems of context-aware coordination introduced by mobility and data heterogeneity in mobile agent systems. Furthermore, environment specific and application specific coordination policy can be integrated into the programmable tuple space for customized requirements. Finally an application sample system-information retrieval in mobile agent applications is carried out to test the performance of the proposed model.

  8. Design and implementation of space physics multi-model application integration based on web

    Science.gov (United States)

    Jiang, Wenping; Zou, Ziming

    With the development of research on space environment and space science, how to develop network online computing environment of space weather, space environment and space physics models for Chinese scientific community is becoming more and more important in recent years. Currently, There are two software modes on space physics multi-model application integrated system (SPMAIS) such as C/S and B/S. the C/S mode which is traditional and stand-alone, demands a team or workshop from many disciplines and specialties to build their own multi-model application integrated system, that requires the client must be deployed in different physical regions when user visits the integrated system. Thus, this requirement brings two shortcomings: reducing the efficiency of researchers who use the models to compute; inconvenience of accessing the data. Therefore, it is necessary to create a shared network resource access environment which could help users to visit the computing resources of space physics models through the terminal quickly for conducting space science research and forecasting spatial environment. The SPMAIS develops high-performance, first-principles in B/S mode based on computational models of the space environment and uses these models to predict "Space Weather", to understand space mission data and to further our understanding of the solar system. the main goal of space physics multi-model application integration system (SPMAIS) is to provide an easily and convenient user-driven online models operating environment. up to now, the SPMAIS have contained dozens of space environment models , including international AP8/AE8、IGRF、T96 models,and solar proton prediction model、geomagnetic transmission model,etc. which are developed by Chinese scientists. another function of SPMAIS is to integrate space observation data sets which offers input data for models online high-speed computing. In this paper, service-oriented architecture (SOA) concept that divides

  9. Separated representations and PGD-based model reduction fundamentals and applications

    CERN Document Server

    Ladevèze, Pierre

    2014-01-01

    The papers in this volume start with a description of  the construction of reduced models through a review of Proper Orthogonal Decomposition (POD) and reduced basis models, including their mathematical foundations and some challenging applications, then followed by a description of a  new generation of simulation strategies based on the use of separated representations (space-parameters, space-time, space-time-parameters, space-space,…), which have led to what is known as Proper Generalized Decomposition (PGD) techniques. The models can be enriched by treating parameters as additional coordinates, leading to fast and inexpensive online calculations based on richer offline parametric solutions. Separated representations are analyzed in detail in the course, from their mathematical foundations to their most spectacular applications. It is also shown how such an approximation could evolve into a new paradigm in computational science, enabling one to circumvent various computational issues in a vast array of...

  10. A bootstrap based space-time surveillance model with an application to crime occurrences

    Science.gov (United States)

    Kim, Youngho; O'Kelly, Morton

    2008-06-01

    This study proposes a bootstrap-based space-time surveillance model. Designed to find emerging hotspots in near-real time, the bootstrap based model is characterized by its use of past occurrence information and bootstrap permutations. Many existing space-time surveillance methods, using population at risk data to generate expected values, have resulting hotspots bounded by administrative area units and are of limited use for near-real time applications because of the population data needed. However, this study generates expected values for local hotspots from past occurrences rather than population at risk. Also, bootstrap permutations of previous occurrences are used for significant tests. Consequently, the bootstrap-based model, without the requirement of population at risk data, (1) is free from administrative area restriction, (2) enables more frequent surveillance for continuously updated registry database, and (3) is readily applicable to criminology and epidemiology surveillance. The bootstrap-based model performs better for space-time surveillance than the space-time scan statistic. This is shown by means of simulations and an application to residential crime occurrences in Columbus, OH, year 2000.

  11. Integrated knowledge-based modeling and its application for classification problems

    Institute of Scientific and Technical Information of China (English)

    2008-01-01

    Knowledge discovery from data directly can hardly avoid the fact that it is biased towards the collected experimental data, whereas, expert systems are always baffled with the manual knowledge acquisition bottleneck. So it is believable that integrating the knowledge embedded in data and those possessed by experts can lead to a superior modeling approach. Aiming at the classification problems, a novel integrated knowledge-based modeling methodology, oriented by experts and driven by data, is proposed. It starts from experts identifying modeling parameters, and then the input space is partitioned followed by fuzzification. Afterwards, single rules are generated and then aggregated to form a rule base. on which a fuzzy inference mechanism is proposed. The experts are allowed to make necessary changes on the rule base to improve the model accuracy. A real-world application, welding fault diagnosis, is presented to demonstrate the effectiveness of the methodology.

  12. Application of Holdridge life-zone model based on the terrain factor in Xinjiang Automous Region

    Institute of Scientific and Technical Information of China (English)

    NI Yong-ming; OUYANG Zhi-yun; WANG Xiao-ke

    2005-01-01

    This study improved the application of the Holdridge life-zone model to simulate the distribution of desert vegetation in China which gives statistics to support eco-recovery and ecosystem reconstruction in desert area. This study classified the desert vegetation into four types: (1) LAD: little arbor desert; (2) SD: shrub desert; (3) HLHSD: half-shrub, little half-shrub desert; (4) LHSCD: little halfshrub cushion desert. Based on the classification of Xinjiang desert vegetation, the classical Holdridge life-zone model was used to simulate Xinjiang desert vegetation's distribution and compare the Kappa coefficient result of the model with table of accuracy represented by Kappa values. The Kappa value of the model was only 0.19, it means the simulation result was poor. To improve the life-zone model application to Xinjiang desert vegetation type, a set of plot standards for terrain factors was developed by using the plot standard as the reclassification criterion to climate sub-regime. Then the desert vegetation in Xinjiang was simulated. The average Kappa value of the second simulation to the respective climate regime was 0.45. The Kappa value of final modeling result was 0.64, which is the better value.The modification of the model made it in more application region. In the end, the model' s ecological relevance to the Xinjiang desert vegetation types was studied.

  13. An Architectural Model for a Grid based Workflow Management Platform in Scientific Applications

    CERN Document Server

    Costan, Alexandru; Stratan, Corina; Dobre, Ciprian; Leordeanu, Catalin; Cristea, Valentin

    2011-01-01

    With recent increasing computational and data requirements of scientific applications, the use of large clustered systems as well as distributed resources is inevitable. Although executing large applications in these environments brings increased performance, the automation of the process becomes more and more challenging. While the use of complex workflow management systems has been a viable solution for this automation process in business oriented environments, the open source engines available for scientific applications lack some functionalities or are too difficult to use for non-specialists. In this work we propose an architectural model for a grid based workflow management platform providing features like an intuitive way to describe workflows, efficient data handling mechanisms and flexible fault tolerance support. Our integrated solution introduces a workflow engine component based on ActiveBPEL extended with additional functionalities and a scheduling component providing efficient mapping between ta...

  14. Artificial neural networks: Principle and application to model based control of drying systems -- A review

    Energy Technology Data Exchange (ETDEWEB)

    Thyagarajan, T.; Ponnavaikko, M. [Crescent Engineering Coll., Madras (India); Shanmugam, J. [Madras Inst. of Tech. (India); Panda, R.C.; Rao, P.G. [Central Leather Research Inst., Madras (India)

    1998-07-01

    This paper reviews the developments in the model based control of drying systems using Artificial Neural Networks (ANNs). Survey of current research works reveals the growing interest in the application of ANN in modeling and control of non-linear, dynamic and time-variant systems. Over 115 articles published in this area are reviewed. All landmark papers are systematically classified in chronological order, in three distinct categories; namely, conventional feedback controllers, model based controllers using conventional methods and model based controllers using ANN for drying process. The principles of ANN are presented in detail. The problems and issues of the drying system and the features of various ANN models are dealt with up-to-date. ANN based controllers lead to smoother controller outputs, which would increase actuator life. The paper concludes with suggestions for improving the existing modeling techniques as applied to predicting the performance characteristics of dryers. The hybridization techniques, namely, neural with fuzzy logic and genetic algorithms, presented, provide, directions for pursuing further research for the implementation of appropriate control strategies. The authors opine that the information presented here would be highly beneficial for pursuing research in modeling and control of drying process using ANN. 118 refs.

  15. Space vector-based modeling and control of a modular multilevel converter in HVDC applications

    DEFF Research Database (Denmark)

    Bonavoglia, M.; Casadei, G.; Zarri, L.; Mengoni, M.; Tani, A.; Serra, G.; Teodorescu, Remus

    Modular multilevel converter (MMC) is an emerging multilevel topology for high-voltage applications that has been developed in recent years. In this paper, the modeling and the control of MMCs are restated in terms of space vectors, which may allow a deeper understanding of the converter behavior....... As a result, a control scheme for three-phase MMCs based on the previous theoretical analysis is presented. Numerical simulations are used to test its feasibility....

  16. Development of Design Procedures for Flexural Applications of Textile Composite Systems Based on Tension Stiffening Models

    OpenAIRE

    Mobasher, Barzin

    2011-01-01

    The Aveston Copper and Kelly (ACK) Method has been routinely used in estimating the efficiency of the bond between the textile and cementitious matrix. This method however has a limited applicability due to the simplifying assumptions such as perfect bond. A numerical model for simulation of tensile behavior of reinforced cement-based composites is presented to capture the inefficiency of the bond mechanisms. In this approach the role of interface properties which are instrumental in the simu...

  17. Application of a CFD based containment model to different large-scale hydrogen distribution experiments

    International Nuclear Information System (INIS)

    Highlights: • A CFD based model developed in ANSYS-FLUENT for simulating the distribution of hydrogen in the containment of a nuclear power plant during a severe accident is validated against four large-scale experiments. • The successive formation and mixing of a stratified gas-layer in experiments performed in the THAI and PANDA facilities are predicted well by the CFD model. • The pressure evolution and related condensation rate during different mixed convection flow conditions in the TOSQAN facility are predicted well by the CFD model. • The results give confidence in the general applicability of the CFD model and model settings. - Abstract: In the event of core degradation during a severe accident in water-cooled nuclear power plants (NPPs), large amounts of hydrogen are generated that may be released into the reactor containment. As the hydrogen mixes with the air in the containment, it can form a flammable mixture. Upon ignition it can damage relevant safety systems and put the integrity of the containment at risk. Despite the installation of mitigation measures, it has been recognized that the temporary existence of combustible or explosive gas clouds cannot be fully excluded during certain postulated accident scenarios. The distribution of hydrogen in the containment and mitigation of the risk are, therefore, important safety issues for NPPs. Complementary to lumped parameter code modelling, Computational Fluid Dynamics (CFD) modelling is needed for the detailed assessment of the hydrogen risk in the containment and for the optimal design of hydrogen mitigation systems in order to reduce this risk as far as possible. The CFD model applied by NRG makes use of the well-developed basic features of the commercial CFD package ANSYS-FLUENT. This general purpose CFD package is complemented with specific user-defined sub-models required to capture the relevant thermal-hydraulic phenomena in the containment during a severe accident as well as the effect of

  18. An emission source inversion model based on satellite data and its application in air quality forecasts

    Institute of Scientific and Technical Information of China (English)

    2010-01-01

    This paper aims at constructing an emission source inversion model using a variational processing method and adaptive nudging scheme for the Community Multiscale Air Quality Model (CMAQ) based on satellite data to investigate the applicability of high resolution OMI (Ozone Monitoring Instrument) column concentration data for air quality forecasts over the North China. The results show a reasonable consistency and good correlation between the spatial distributions of NO2 from surface and OMI satellite measurements in both winter and summer. Such OMI products may be used to implement integrated variational analysis based on observation data on the ground. With linear and variational corrections made, the spatial distribution of OMI NO2 clearly revealed more localized distributing characteristics of NO2 concentration. With such information, emission sources in the southwest and southeast of North China are found to have greater impacts on air quality in Beijing. When the retrieved emission source inventory based on high-resolution OMI NO2 data was used, the coupled Weather Research Forecasting CMAQ model (WRF-CMAQ) performed significantly better in forecasting NO2 concentration level and its tendency as reflected by the more consistencies between the NO2 concentrations from surface observation and model result. In conclusion, satellite data are particularly important for simulating NO2 concentrations on urban and street-block scale. High-resolution OMI NO2 data are applicable for inversing NOx emission source inventory, assessing the regional pollution status and pollution control strategy, and improving the model forecasting results on urban scale.

  19. A New Web-based Application Optimization Model in Multicore Web Server

    Science.gov (United States)

    You, Guohua; Zhao, Ying

    More and more web servers adopt multi-core CPUs to improve performance because of the development of multi-core technology. However, web applications couldn't exploit the potential of multi-core web server efficiently because of traditional processing algorithm of requests and scheduling strategies of threads in O/S. In this paper, a new web-based application optimization model was proposed, which could classify and schedule the dynamic requests and static requests on scheduling core, and process the dynamic requests on the other cores. By this way, a simulation program, which is called SIM, was developed. Experiments have been done to validate the new model, and the results show that the new model can effectively improve the performance of multi-core web servers, and avoid the problems of ping-pong effect.

  20. GRace: a MATLAB-based application for fitting the discrimination-association model.

    Science.gov (United States)

    Stefanutti, Luca; Vianello, Michelangelo; Anselmi, Pasquale; Robusto, Egidio

    2014-01-01

    The Implicit Association Test (IAT) is a computerized two-choice discrimination task in which stimuli have to be categorized as belonging to target categories or attribute categories by pressing, as quickly and accurately as possible, one of two response keys. The discrimination association model has been recently proposed for the analysis of reaction time and accuracy of an individual respondent to the IAT. The model disentangles the influences of three qualitatively different components on the responses to the IAT: stimuli discrimination, automatic association, and termination criterion. The article presents General Race (GRace), a MATLAB-based application for fitting the discrimination association model to IAT data. GRace has been developed for Windows as a standalone application. It is user-friendly and does not require any programming experience. The use of GRace is illustrated on the data of a Coca Cola-Pepsi Cola IAT, and the results of the analysis are interpreted and discussed. PMID:26054728

  1. Model Test Based Soil Spring Model and Application in Pipeline Thermal Buckling Analysis

    Institute of Scientific and Technical Information of China (English)

    GAO Xi-feng; LIU Run; YAN Shu-wang

    2011-01-01

    The buckling of submarine pipelines may occur due to the action of axial soil frictional force caused by relative movement of soil and pipeline,which is induced by the thermal and internal pressure.The likelihood of occurrence of this buckling phenomenon is largely determined by soil resistance.A series of large-scale model tests were carried out to facilitate the establishment of substantial data base for a variety of burial pipeline relationships.Based on the test data,nonlinear soil spring can be adopted to simulate the soil behavior during the pipeline movement.For uplift resistance,an ideal elasticity plasticity model is recommended in the case of H/D (depth-to-diameter ratio)>5 and an elasticity softened model is recommended in the case of H/D≤5.The soil resistance along the pipeline axial direction can be simulated by an ideal elasticity plasticity model.The numerical analyzing results show that the capacity of pipeline against thermal buckling decreases with its initial imperfection enlargement and increases with the burial depth enhancement.

  2. Combined model- and rule-based controller synthesis with application to helicopter flight

    Science.gov (United States)

    Jiang, Tian-Yue

    This thesis deals with synthesis of combined (nonlinear) model-based and (fuzzy logic) rule-based controllers, along with their applications to helicopter flight control problem. The synthesis involves superimposing two control techniques in order to meet both stability and performance objectives. One is model-based control technique, which is based on inversion of an approximate model of the real system. The other is rule-based control technique that adaptively cancels the inversion errors caused by the approximate model inversion. There are two major aspects of the research effort in this thesis. The first is the development of the adaptive rule-based (fuzzy logic) controllers. The linguistic rule weights and defuzzification output weights in the controllers are adapted for ultimate boundedness of the tracking errors. Numerical results from a helicopter flight control problem indicate improvement and demonstrate effectiveness of the control technique. The second aspect of this research work is the extension of the synthesis to account for control limits. In this thesis, a control saturation related rule-bank in conjunction with the adaptive fuzzy logic controller is designed to trade-off system performance for closed-loop stability when the tendency towards control amplitude and/or rate saturation is detected. Simulation results from both a fixed-wing aircraft trajectory control problem and a helicopter flight control problem show the effectiveness of the synthesis method and the resulting controller in avoiding control saturations.

  3. Application of recently developed elliptic blending based models to separated flows

    International Nuclear Information System (INIS)

    Highlights: ► The study focuses on elliptic blending near-wall models. ► Models are compared on 2- and 3-dimensional separating flows. ► Conclusions are ambiguous on 2-d flows. ► Predictive superiority of Reynolds stress models over eddy viscosity model appear on 3-d flows. - Abstract: This paper considers the application of four Reynolds-Averaged Navier Stokes (RANS) models to a range of progressively complex test cases, exhibiting both 2-d and 3-d flow separation. Two Eddy Viscosity Models (EVM) and two Reynolds Stress Transport Models (RSM) are employed, of which two (one in each category) are based on elliptic blending formulations. By both reviewing the conclusions of previous studies, and from the present calculations, this study aims at gaining more insight into the importance of two modelling features for these flows: the usage of turbulence anisotropy resolving schemes, and the near-wall limiting behaviour. In general the anisotropy and near wall treatment offered by both elliptic blending models is observed to offer some improvement over other models tested, although this is not always the case for the 2-d flows, where (as ever) a single “best candidate” model does not emerge.

  4. Constructing a raster-based spatio-temporal hierarchical data model for marine fisheries application

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    Marine information has been increasing quickly. The traditional database technologies have disadvantages in manipulating large amounts of marine information which relates to the position in 3-D with the time. Recently, greater emphasis has been placed on GIS (geographical information system)to deal with the marine information. The GIS has shown great success for terrestrial applications in the last decades, but its use in marine fields has been far more restricted. One of the main reasons is that most of the GIS systems or their data models are designed for land applications. They cannot do well with the nature of the marine environment and for the marine information. And this becomes a fundamental challenge to the traditional GIS and its data structure. This work designed a data model,the raster-based spatio-temporal hierarchical data model (RSHDM), for the marine information system, or for the knowledge discovery from spatio-temporal data, which bases itself on the nature of the marine data and overcomes the shortages of the current spatio-temporal models when they are used in the field. As an experiment, the marine fishery data warehouse (FDW) for marine fishery management was set up, which was based on the RSHDM. The experiment proved that the RSHDM can do well with the data and can extract easily the aggregations that the management needs at different levels.

  5. Development of web-based reliability data analysis algorithm model and its application

    International Nuclear Information System (INIS)

    For this study, a database model of plant reliability was developed for the effective acquisition and management of plant-specific data that can be used in various applications of plant programs as well as in Probabilistic Safety Assessment (PSA). Through the development of a web-based reliability data analysis algorithm, this approach systematically gathers specific plant data such as component failure history, maintenance history, and shift diary. First, for the application of the developed algorithm, this study reestablished the raw data types, data deposition procedures and features of the Enterprise Resource Planning (ERP) system process. The component codes and system codes were standardized to make statistical analysis between different types of plants possible. This standardization contributes to the establishment of a flexible database model that allows the customization of reliability data for the various applications depending on component types and systems. In addition, this approach makes it possible for users to perform trend analyses and data comparisons for the significant plant components and systems. The validation of the algorithm is performed through a comparison of the importance measure value (Fussel-Vesely) of the mathematical calculation and that of the algorithm application. The development of a reliability database algorithm is one of the best approaches for providing systemic management of plant-specific reliability data with transparency and continuity. This proposed algorithm reinforces the relationships between raw data and application results so that it can provide a comprehensive database that offers everything from basic plant-related data to final customized data.

  6. A Stress Vector-Based Constitutive Model for Cohesionless Soil( Ⅱ )-Application

    Institute of Scientific and Technical Information of China (English)

    史宏彦; 谢定义; 白琳

    2002-01-01

    The stress vector-based constitutive model for cohesionless soil, proposed by SHI Hong-yan et al., was applied to analyze the deformation behaviors of materials subjected to various stress paths. The result of analysis shows that the constitutive model can capture well the main deformation behavior of cohesionless soil, such as stress-strain nonlinearity,hardening property, dilatancy , stress path dependency, non- coaxiality between the principal stress and the principal strain increment directions, and the coupling of mean effective and deviatoric stress with deformation. In addition, the model can also take into account the rotation of principal stress axes and the influence of intermediate principal stress on deformation and strength of soil simultaneously. The excellent agreement between the predicted and measured behavior indicates the comprehensive applicability of the model.

  7. Copula based prediction models: an application to an aortic regurgitation study

    Directory of Open Access Journals (Sweden)

    Shoukri Mohamed M

    2007-06-01

    Full Text Available Abstract Background: An important issue in prediction modeling of multivariate data is the measure of dependence structure. The use of Pearson's correlation as a dependence measure has several pitfalls and hence application of regression prediction models based on this correlation may not be an appropriate methodology. As an alternative, a copula based methodology for prediction modeling and an algorithm to simulate data are proposed. Methods: The method consists of introducing copulas as an alternative to the correlation coefficient commonly used as a measure of dependence. An algorithm based on the marginal distributions of random variables is applied to construct the Archimedean copulas. Monte Carlo simulations are carried out to replicate datasets, estimate prediction model parameters and validate them using Lin's concordance measure. Results: We have carried out a correlation-based regression analysis on data from 20 patients aged 17–82 years on pre-operative and post-operative ejection fractions after surgery and estimated the prediction model: Post-operative ejection fraction = - 0.0658 + 0.8403 (Pre-operative ejection fraction; p = 0.0008; 95% confidence interval of the slope coefficient (0.3998, 1.2808. From the exploratory data analysis, it is noted that both the pre-operative and post-operative ejection fractions measurements have slight departures from symmetry and are skewed to the left. It is also noted that the measurements tend to be widely spread and have shorter tails compared to normal distribution. Therefore predictions made from the correlation-based model corresponding to the pre-operative ejection fraction measurements in the lower range may not be accurate. Further it is found that the best approximated marginal distributions of pre-operative and post-operative ejection fractions (using q-q plots are gamma distributions. The copula based prediction model is estimated as: Post -operative ejection fraction = - 0.0933 + 0

  8. Physiologically based pharmacokinetic modeling using microsoft excel and visual basic for applications.

    Science.gov (United States)

    Marino, Dale J

    2005-01-01

    Abstract Physiologically based pharmacokinetic (PBPK) models are mathematical descriptions depicting the relationship between external exposure and internal dose. These models have found great utility for interspecies extrapolation. However, specialized computer software packages, which are not widely distributed, have typically been used for model development and utilization. A few physiological models have been reported using more widely available software packages (e.g., Microsoft Excel), but these tend to include less complex processes and dose metrics. To ascertain the capability of Microsoft Excel and Visual Basis for Applications (VBA) for PBPK modeling, models for styrene, vinyl chloride, and methylene chloride were coded in Advanced Continuous Simulation Language (ACSL), Excel, and VBA, and simulation results were compared. For styrene, differences between ACSL and Excel or VBA compartment concentrations and rates of change were less than +/-7.5E-10 using the same numerical integration technique and time step. Differences using VBA fixed step or ACSL Gear's methods were generally Excel and VBA PBPK model dose metrics differed by no more than -0.013% or -0.23%, respectively, from ACSL results. These differences are likely attributable to different step sizes rather than different numerical integration techniques. These results indicate that Microsoft Excel and VBA can be useful tools for utilizing PBPK models, and given the availability of these software programs, it is hoped that this effort will help facilitate the use and investigation of PBPK modeling. PMID:20021074

  9. Web-based Services for Earth Observing and Model Data in National Applications and Hazards

    Science.gov (United States)

    Kafatos, M.; Boybeyi, Z.; Cervone, G.; di, L.; Sun, D.; Yang, C.; Yang, R.

    2005-12-01

    The ever-growing large volumes of Earth system science data, collected by Earth observing platforms, in situ stations and as model output data, are increasingly being used by discipline scientists and by wider classes of users. In particular, applications of Earth system science data to environmental and hazards as well as other national applications, require tailored or specialized data, as well as web-based tools and infrastructure. The latter are driven by applications and usage drivers which include ease of access, visualization of complex data, ease of producing value-added data, GIS and open source analysis usage, metadata, etc. Here we present different aspects of such web-based services and access, and discuss several applications in the hazards and environmental areas, including earthquake signatures and observations and model runs of hurricanes. Examples and lessons learned from the consortium Mid-Atlantic Geospatial Information Consortium will be presented. We discuss a NASA-funded, open source on-line data analysis system that is being applied to climate studies for the ESIP Federation. Since enhanced, this project and the next-generation Metadata Integrated Data Analysis System allow users not only to identify data but also to generate new data products on-the-fly. The functionalities extend from limited predefined functions, to sophisticated functions described by general-purposed GrADS (Grid Analysis and Display System) commands. The Federation system also allows third party data products to be combined with local data. Software component are available for converting the output from MIDAS (OPenDAP) into OGC compatible software. The on-going Grid efforts at CEOSR and LAITS in the School of Computational Sciences (SCS) include enhancing the functions of Globus to provide support for a geospatial system so the system can share the computing power to handle problems with different peak access times and improve the stability and flexibility of a rapid

  10. Criterion of applicable models for planar type Cherenkov laser based on quantum mechanical treatments

    International Nuclear Information System (INIS)

    A generalized theoretical analysis for amplification mechanism in the planar-type Cherenkov laser is given. An electron is represented to be a material wave having temporal and spatial varying phases with finite spreading length. Interaction between the electrons and the electromagnetic (EM) wave is analyzed by counting the quantum statistical properties. The interaction mechanism is classified into the Velocity and Density Modulation (VDM) model and the Energy Level Transition (ELT) model basing on the relation between the wavelength of the EM wave and the electron spreading length. The VDM model is applicable when the wavelength of the EM wave is longer than the electron spreading length as in the microwave region. The dynamic equation of the electron, which is popularly used in the classical Newtonian mechanics, has been derived from the quantum mechanical Schrödinger equation. The amplification of the EM wave can be explained basing on the bunching effect of the electron density in the electron beam. The amplification gain and whose dispersion relation with respect to the electron velocity is given in this paper. On the other hand, the ELT model is applicable for the case that the wavelength of the EM wave is shorter than the electron spreading length as in the optical region. The dynamics of the electron is explained to be caused by the electron transition between different energy levels. The amplification gain and whose dispersion relation with respect to the electron acceleration voltage was derived on the basis of the quantum mechanical density matrix

  11. Application of a dislocation based model for Interstitial Free (IF) steels to typical stamping simulations

    Science.gov (United States)

    Carvalho Resende, T.; Balan, T.; Abed-Meraim, F.; Bouvier, S.; Sablin, S.-S.

    2010-06-01

    With a view to environmental, economic and safety concerns, car manufacturers need to design lighter and safer vehicles in ever shorter development times. In recent years, High Strength Steels (HSS) like Interstitial Free (IF) steels which have higher ratios of yield strength to elastic modulus, are increasingly used for sheet metal parts in automotive industry to meet the demands. Moreover, the application of sheet metal forming simulations has proven to be beneficial to reduce tool costs in the design stage and to optimize current processes. The Finite Element Method (FEM) is quite successful to simulate metal forming processes but accuracy largely depends on the quality of the material properties provided as input to the material model. Common phenomenological models roughly consist in the fitting of functions on experimental results and do not provide any predictive character for different metals from the same grade. Therefore, the use of accurate plasticity models based on physics would increase predictive capability, reduce parameter identification cost and allow for robust and time-effective finite element simulations. For this purpose, a 3D physically based model at large strain with dislocation density evolution approach was presented in IDDRG2009 by the authors [1]. This model allows the description of work-hardening's behavior for different loading paths (i.e. uni-axial tensile, simple shear and Bauschinger tests) taking into account several data from microstructure (i.e. grain size, texture, etc…). The originality of this model consists in the introduction of microstructure data in a classical phenomenological model in order to achieve work-hardening's predictive character for different metals from the same grade. Indeed, thanks to a microstructure parameter set for an Interstitial Free steel, it is possible to describe work-hardening behavior for different loading paths of other IF steels by only changing the mean grain size and the chemical

  12. On the limits of applicability of drift-diffusion based hot carrier degradation modeling

    Science.gov (United States)

    Jech, Markus; Sharma, Prateek; Tyaginov, Stanislav; Rudolf, Florian; Grasser, Tibor

    2016-04-01

    We study the limits of the applicability of a drift-diffusion (DD) based model for hot-carrier degradation (HCD). In this approach the rigorous but computationally expensive solution of the Boltzmann transport equation is replaced by an analytic expression for the carrier energy distribution function. On the one hand, we already showed that the simplified version of our HCD model is quite successful for LDMOS devices. On the other hand, hot carrier degradation models based on the drift-diffusion and energy transport schemes were shown to fail for planar MOSFETs with gate lengths of 0.5-2.0 µm. To investigate the limits of validity of the DD-based HCD model, we use planar nMOSFETs of an identical topology but with different gate lengths of 2.0, 1.5, and 1.0 µm. We show that, although the model is able to adequately represent the linear and saturation drain current changes in the 2.0 µm transistor, it starts to fail for gate lengths shorter than 1.5 µm and becomes completely inadequate for the 1.0 µm device.

  13. Application of Levenberg-Marquardt Optimization Algorithm Based Multilayer Neural Networks for Hydrological Time Series Modeling

    Directory of Open Access Journals (Sweden)

    Umut Okkan

    2011-07-01

    Full Text Available Recently, Artificial Neural Networks (ANN, which is mathematical modelingtools inspired by the properties of the biological neural system, has been typically used inthe studies of hydrological time series modeling. These modeling studies generally includethe standart feed forward backpropagation (FFBP algorithms such as gradient-descent,gradient-descent with momentum rate and, conjugate gradient etc. As the standart FFBPalgorithms have some disadvantages relating to the time requirement and slowconvergency in training, Newton and Levenberg-Marquardt algorithms, which arealternative approaches to standart FFBP algorithms, were improved and also used in theapplications. In this study, an application of Levenberg-Marquardt algorithm based ANN(LM-ANN for the modeling of monthly inflows of Demirkopru Dam, which is located inthe Gediz basin, was presented. The LM-ANN results were also compared with gradientdescentwith momentum rate algorithm based FFBP model (GDM-ANN. When thestatistics of the long-term and also seasonal-term outputs are compared, it can be seen thatthe LM-ANN model that has been developed, is more sensitive for prediction of theinflows. In addition, LM-ANN approach can be used for modeling of other hydrologicalcomponents in terms of a rapid assessment and its robustness.

  14. Model-based geostatistics

    CERN Document Server

    Diggle, Peter J

    2007-01-01

    Model-based geostatistics refers to the application of general statistical principles of modeling and inference to geostatistical problems. This volume provides a treatment of model-based geostatistics and emphasizes on statistical methods and applications. It also features analyses of datasets from a range of scientific contexts.

  15. UNDERSTANDING THE APPLICABILITY OF LINEAR & NON-LINEAR MODELS USING A CASE-BASED STUDY

    OpenAIRE

    Gaurav Singh Thakur; Anubhav Gupta; Ankur Bhardwaj; Biju R Mohan

    2014-01-01

    This paper uses a case based study – “product sales estimation” on real-time data to help us understand the applicability of linear and non-linear models in machine learning and data mining. A systematic approach has been used here to address the given problem statement of sales estimation for a particular set of products in multiple categories by applying both linear and non-linear machine learning techniques on a data set of selected features from the original data set. Feature ...

  16. Establishment of Winter Wheat Regional Simulation Model Based on Remote Sensing Data and Its Application

    Institute of Scientific and Technical Information of China (English)

    MA Yuping; WANG Shili; ZHANG Li; HOU Yingyu; ZHUANG Liwei; WANG Futang

    2006-01-01

    Accurate crop growth monitoring and yield forecasting are significant to the food security and the sus tainable development of agriculture. Crop yield estimation by remote sensing and crop growth simulation models have highly potential application in crop growth monitoring and yield forecasting. However, both of them have limitations in mechanism and regional application, respectively. Therefore, approach and methodology study on the combination of remote sensing data and crop growth simulation models are con cerned by many researchers. In this paper, adjusted and regionalized WOFOST (World Food Study) in North China and Scattering by Arbitrarily Inclined Leaves-a model of leaf optical PROperties SPECTra (SAIL-PROSFPECT) were coupled through LAI to simulate Soil Adjusted Vegetation Index (SAVI) of crop canopy, by which crop model was re-initialized by minimizing differences between simulated and synthesized SAVI from remote sensing data using an optimization software (FSEOPT). Thus, a regional remote-sensing crop-simulation-framework-model (WSPFRS) was established under potential production level (optimal soil water condition). The results were as follows: after re-initializing regional emergence date by using remote sensing data, anthesis, and maturity dates simulated by WSPFRS model were more close to measured values than simulated results of WOFOST; by re-initializing regional biomass weight at turn-green stage, the spa tial distribution of simulated storage organ weight was more consistent with measured yields and the area with high values was nearly consistent with actual high yield area. This research is a basis for developing regional crop model in water stress production level based on remote sensing data.

  17. Development and Application of a Process-based River System Model at a Continental Scale

    Science.gov (United States)

    Kim, S. S. H.; Dutta, D.; Vaze, J.; Hughes, J. D.; Yang, A.; Teng, J.

    2014-12-01

    Existing global and continental scale river models, mainly designed for integrating with global climate model, are of very course spatial resolutions and they lack many important hydrological processes, such as overbank flow, irrigation diversion, groundwater seepage/recharge, which operate at a much finer resolution. Thus, these models are not suitable for producing streamflow forecast at fine spatial resolution and water accounts at sub-catchment levels, which are important for water resources planning and management at regional and national scale. A large-scale river system model has been developed and implemented for water accounting in Australia as part of the Water Information Research and Development Alliance between Australia's Bureau of Meteorology (BoM) and CSIRO. The model, developed using node-link architecture, includes all major hydrological processes, anthropogenic water utilisation and storage routing that influence the streamflow in both regulated and unregulated river systems. It includes an irrigation model to compute water diversion for irrigation use and associated fluxes and stores and a storage-based floodplain inundation model to compute overbank flow from river to floodplain and associated floodplain fluxes and stores. An auto-calibration tool has been built within the modelling system to automatically calibrate the model in large river systems using Shuffled Complex Evolution optimiser and user-defined objective functions. The auto-calibration tool makes the model computationally efficient and practical for large basin applications. The model has been implemented in several large basins in Australia including the Murray-Darling Basin, covering more than 2 million km2. The results of calibration and validation of the model shows highly satisfactory performance. The model has been operalisationalised in BoM for producing various fluxes and stores for national water accounting. This paper introduces this newly developed river system model

  18. Interdependent multi-layer networks: modeling and survivability analysis with applications to space-based networks.

    Directory of Open Access Journals (Sweden)

    Jean-Francois Castet

    Full Text Available This article develops a novel approach and algorithmic tools for the modeling and survivability analysis of networks with heterogeneous nodes, and examines their application to space-based networks. Space-based networks (SBNs allow the sharing of spacecraft on-orbit resources, such as data storage, processing, and downlink. Each spacecraft in the network can have different subsystem composition and functionality, thus resulting in node heterogeneity. Most traditional survivability analyses of networks assume node homogeneity and as a result, are not suited for the analysis of SBNs. This work proposes that heterogeneous networks can be modeled as interdependent multi-layer networks, which enables their survivability analysis. The multi-layer aspect captures the breakdown of the network according to common functionalities across the different nodes, and it allows the emergence of homogeneous sub-networks, while the interdependency aspect constrains the network to capture the physical characteristics of each node. Definitions of primitives of failure propagation are devised. Formal characterization of interdependent multi-layer networks, as well as algorithmic tools for the analysis of failure propagation across the network are developed and illustrated with space applications. The SBN applications considered consist of several networked spacecraft that can tap into each other's Command and Data Handling subsystem, in case of failure of its own, including the Telemetry, Tracking and Command, the Control Processor, and the Data Handling sub-subsystems. Various design insights are derived and discussed, and the capability to perform trade-space analysis with the proposed approach for various network characteristics is indicated. The select results here shown quantify the incremental survivability gains (with respect to a particular class of threats of the SBN over the traditional monolith spacecraft. Failure of the connectivity between nodes is also

  19. Comparing GIS-based habitat models for applications in EIA and SEA

    International Nuclear Information System (INIS)

    Land use changes, urbanisation and infrastructure developments in particular, cause fragmentation of natural habitats and threaten biodiversity. Tools and measures must be adapted to assess and remedy the potential effects on biodiversity caused by human activities and developments. Within physical planning, environmental impact assessment (EIA) and strategic environmental assessment (SEA) play important roles in the prediction and assessment of biodiversity-related impacts from planned developments. However, adapted prediction tools to forecast and quantify potential impacts on biodiversity components are lacking. This study tested and compared four different GIS-based habitat models and assessed their relevance for applications in environmental assessment. The models were implemented in the Stockholm region in central Sweden and applied to data on the crested tit (Parus cristatus), a sedentary bird species of coniferous forest. All four models performed well and allowed the distribution of suitable habitats for the crested tit in the Stockholm region to be predicted. The models were also used to predict and quantify habitat loss for two regional development scenarios. The study highlighted the importance of model selection in impact prediction. Criteria that are relevant for the choice of model for predicting impacts on biodiversity were identified and discussed. Finally, the importance of environmental assessment for the preservation of biodiversity within the general frame of biodiversity conservation is emphasised.

  20. A constriction resistance model of conjugated polymer based piezoresistive sensors for electronic skin applications.

    Science.gov (United States)

    Khalili, N; Naguib, H E; Kwon, R H

    2016-05-14

    Human intervention can be replaced through the development of tools resulting from utilization of sensing devices possessing a wide range of applications including humanoid robots or remote and minimally invasive surgeries. Similar to the five human senses, sensors interface with their surroundings to stimulate a suitable response or action. The sense of touch which arises in human skin is among the most challenging senses to emulate due to its ultra high sensitivity. This has brought forth novel challenging issues to consider in the field of biomimetic robotics. In this work, using a multiphase reaction, a polypyrrole (PPy) based hydrogel is developed as a resistive type pressure sensor with an intrinsically elastic microstructure stemming from three dimensional hollow spheres. It is shown that the electrical conductivity of the fabricated PPy based piezoresistive sensors is enhanced as a result of adding conductive fillers and therefore, endowing the sensors with a higher sensitivity. A semi-analytical constriction resistance based model accounting for the real contact area between the PPy hydrogel sensors and the electrode along with the dependency of the contact resistance change on the applied load is developed. The model is then solved using a Monte Carlo technique and its corresponding sensitivity is obtained. Comparing the results with their experimental counterparts, the proposed modeling methodology offers a good tracking ability. PMID:27035514

  1. Autonomous Reactor Control Using Model Based Predictive Control for Space Propulsion Applications

    International Nuclear Information System (INIS)

    Reliable reactor control is important to reactor safety, both in terrestrial and space systems. For a space system, where the time for communication to Earth is significant, autonomous control is imperative. Based on feedback from reactor diagnostics, a controller must be able to automatically adjust to changes in reactor temperature and power level to maintain nominal operation without user intervention. Model-based predictive control (MBPC) (Clarke 1994; Morari 1994) is investigated as a potential control methodology for reactor start-up and transient operation in the presence of an external source. Bragg-Sitton and Holloway (2004) assessed the applicability of MBPC to reactor start-up from a cold, zero-power condition in the presence of a time-varying external radiation source, where large fluctuations in the external radiation source can significantly impact a reactor during start-up operations. The MBPC algorithm applied the point kinetics model to describe the reactor dynamics, using a single group of delayed neutrons; initial application considered a fast neutron lifetime (10-3 sec) to simplify calculations during initial controller analysis. The present study will more accurately specify the dynamics of a fast reactor, using a more appropriate fast neutron lifetime (10-7 sec) than in the previous work. Controller stability will also be assessed by carefully considering the dependencies of each component in the defined cost (objective) function and its subsequent effect on the selected 'optimal' control maneuvers

  2. Research on Application of Power System Reliability Model Based on Principal-agent Mechanism

    Directory of Open Access Journals (Sweden)

    Shaomei Yang

    2012-02-01

    Full Text Available Power system reliability has been a widespread concern. With the continuous development of the electricity market, there has been a serious conflict, which between social obligations and necessary requirement. The so-called social obligation is that the power industry must ensure the power system operation reliability; the so-called necessary requirement is that the power industry must solve the power market operation economics. The power system operation reliability is serious challenge, so we must establish a new mechanism adapting to the electricity market, and ensure the power system operation reliability. In this paper, we have researched the model and method based on the principal-agent mechanism from three aspects, which are the reliability principal-agent model starting point, principal-agent model analysis and design, and information transparency level impact analysis. On this basis, we proposed the reliability management model based on the principal-agent mechanism, and carried out example analysis. The example analysis shows that through establishing the principal-agent relationship between power enterprises and the user, the entire power system reliability resources would be more efficient configuration to further enhance the entire society electricity reliability benefits. Further evidence that the principal-agent mechanism is scientific and applicability in the power system reliability study.

  3. Application of uncertainty reasoning based on cloud model in time series prediction

    Institute of Scientific and Technical Information of China (English)

    张锦春; 胡谷雨

    2003-01-01

    Time series prediction has been successfully used in several application areas, such as meteorological forecasting, market prediction, network traffic forecasting, etc., and a number of techniques have been developed for modeling and predicting time series. In the traditional exponential smoothing method, a fixed weight is assigned to data history, and the trend changes of time series are ignored. In this paper, an uncertainty reasoning method, based on cloud model, is employed in time series prediction, which uses cloud logic controller to adjust the smoothing coefficient of the simple exponential smoothing method dynamically to fit the current trend of the time series. The validity of this solution was proved by experiments on various data sets.

  4. Application of uncertainty reasoning based on cloud model in time series prediction

    Institute of Scientific and Technical Information of China (English)

    张锦春; 胡谷雨

    2003-01-01

    Time series prediction has been successfully used in several application areas, such as meteoro-logical forecasting, market prediction, network traffic forecasting, etc. , and a number of techniques have been developed for modeling and predicting time series. In the traditional exponential smoothing method, a fixed weight is assigned to data history, and the trend changes of time series are ignored. In this paper, an uncertainty reasoning method, based on cloud model, is employed in time series prediction, which uses cloud logic controller to adjust the smoothing coefficient of the simple exponential smoothing method dynamically to fit the current trend of the time series. The validity of this solution was proved by experiments on various data sets.

  5. Research on application of intelligent computation based LUCC model in urbanization process

    Science.gov (United States)

    Chen, Zemin

    2007-06-01

    Global change study is an interdisciplinary and comprehensive research activity with international cooperation, arising in 1980s, with the largest scopes. The interaction between land use and cover change, as a research field with the crossing of natural science and social science, has become one of core subjects of global change study as well as the front edge and hot point of it. It is necessary to develop research on land use and cover change in urbanization process and build an analog model of urbanization to carry out description, simulation and analysis on dynamic behaviors in urban development change as well as to understand basic characteristics and rules of urbanization process. This has positive practical and theoretical significance for formulating urban and regional sustainable development strategy. The effect of urbanization on land use and cover change is mainly embodied in the change of quantity structure and space structure of urban space, and LUCC model in urbanization process has been an important research subject of urban geography and urban planning. In this paper, based upon previous research achievements, the writer systematically analyzes the research on land use/cover change in urbanization process with the theories of complexity science research and intelligent computation; builds a model for simulating and forecasting dynamic evolution of urban land use and cover change, on the basis of cellular automation model of complexity science research method and multi-agent theory; expands Markov model, traditional CA model and Agent model, introduces complexity science research theory and intelligent computation theory into LUCC research model to build intelligent computation-based LUCC model for analog research on land use and cover change in urbanization research, and performs case research. The concrete contents are as follows: 1. Complexity of LUCC research in urbanization process. Analyze urbanization process in combination with the contents

  6. A Novel Application of Agent-based Modeling: Projecting Water Access and Availability Using a Coupled Hydrologic Agent-based Model in the Nzoia Basin, Kenya

    Science.gov (United States)

    Le, A.; Pricope, N. G.

    2015-12-01

    Projections indicate that increasing population density, food production, and urbanization in conjunction with changing climate conditions will place stress on water resource availability. As a result, a holistic understanding of current and future water resource distribution is necessary for creating strategies to identify the most sustainable means of accessing this resource. Currently, most water resource management strategies rely on the application of global climate predictions to physically based hydrologic models to understand potential changes in water availability. However, the need to focus on understanding community-level social behaviors that determine individual water usage is becoming increasingly evident, as predictions derived only from hydrologic models cannot accurately represent the coevolution of basin hydrology and human water and land usage. Models that are better equipped to represent the complexity and heterogeneity of human systems and satellite-derived products in place of or in conjunction with historic data significantly improve preexisting hydrologic model accuracy and application outcomes. We used a novel agent-based sociotechnical model that combines the Soil and Water Assessment Tool (SWAT) and Agent Analyst and applied it in the Nzoia Basin, an area in western Kenya that is becoming rapidly urbanized and industrialized. Informed by a combination of satellite-derived products and over 150 household surveys, the combined sociotechnical model provided unique insight into how populations self-organize and make decisions based on water availability. In addition, the model depicted how population organization and current management alter water availability currently and in the future.

  7. UNDERSTANDING THE APPLICABILITY OF LINEAR & NON-LINEAR MODELS USING A CASE-BASED STUDY

    Directory of Open Access Journals (Sweden)

    Gaurav Singh Thakur

    2014-11-01

    Full Text Available This paper uses a case based study – “product sales estimation” on real-time data to help us understand the applicability of linear and non-linear models in machine learning and data mining. A systematic approach has been used here to address the given problem statement of sales estimation for a particular set of products in multiple categories by applying both linear and non-linear machine learning techniques on a data set of selected features from the original data set. Feature selection is a process that reduces the dimensionality of the data set by excluding those features which contribute minimal to the prediction of the dependent variable. The next step in this process is training the model that is done using multiple techniques from linear & non-linear domains, one of the best ones in their respective areas. Data Remodeling has then been done to extract new features from the data set by changing the structure of the dataset & the performance of the models is checked again. Data Remodeling often plays a very crucial and important role in boosting classifier accuracies by changing the properties of the given dataset. We then try to explore and analyze the various reasons due to which one model performs better than the other & hence try and develop an understanding about the applicability of linear & non-linear machine learning models. The target mentioned above being our primary goal, we also aim to find the classifier with the best possible accuracy for product sales estimation in the given scenario.

  8. Kernel based model parametrization and adaptation with applications to battery management systems

    Science.gov (United States)

    Weng, Caihao

    With the wide spread use of energy storage systems, battery state of health (SOH) monitoring has become one of the most crucial challenges in power and energy research, as SOH significantly affects the performance and life cycle of batteries as well as the systems they are interacting with. Identifying the SOH and adapting of the battery energy/power management system accordingly are thus two important challenges for applications such as electric vehicles, smart buildings and hybrid power systems. This dissertation focuses on the identification of lithium ion battery capacity fading, and proposes an on-board implementable model parametrization and adaptation framework for SOH monitoring. Both parametric and non-parametric approaches that are based on kernel functions are explored for the modeling of battery charging data and aging signature extraction. A unified parametric open circuit voltage model is first developed to improve the accuracy of battery state estimation. Several analytical and numerical methods are then investigated for the non-parametric modeling of battery data, among which the support vector regression (SVR) algorithm is shown to be the most robust and consistent approach with respect to data sizes and ranges. For data collected on LiFePO 4 cells, it is shown that the model developed with the SVR approach is able to predict the battery capacity fading with less than 2% error. Moreover, motivated by the initial success of applying kernel based modeling methods for battery SOH monitoring, this dissertation further exploits the parametric SVR representation for real-time battery characterization supported by test data. Through the study of the invariant properties of the support vectors, a kernel based model parametrization and adaptation framework is developed. The high dimensional optimization problem in the learning algorithm could be reformulated as a parameter estimation problem, that can be solved by standard estimation algorithms such as the

  9. A Decision Model for Supplier Selection based on Business System Management and Safety Criteria and Application of the Model

    Directory of Open Access Journals (Sweden)

    Semih Coşkun

    2015-08-01

    Full Text Available In modern market conditions, sustainable and effective management of main manufacturers, suppliers and customer relationship is a necessity for competitiveness. Suppliers must satisfy customers’ expectations such as cost minimization, quality maximization, improved flexibility and achieved deadlines; which is also required for systematic management of products, work and environmental safety. The supplier selection process is consistently getting more complicated by the effect of increasing amount of suppliers and supplier selection criteria. Supplier selection decisions which take an important role in efficient supplier management will be more consistent by the application of decision making models which integrate the quantitative and qualitative evaluation factors. In this study, a dynamic process is designed and modeled for supplier selection. For this purpose, evaluation criteria are established according to the Balanced Scorecard perspectives, system sustainability and safety requirements. Fuzzy Analytic Hierarchy Process method is used for evaluating the importance of supplier selection criteria. A utility range-based interactive group decision making method is used for the selection of the best supplier. In order to test the proposed model, a representative from airport operation sector is selected. Finally, it is revealed that the application of the proposed model generates consistent results for supplier selection decisions.

  10. Model-based Reinforcement Learning with Model Error and Its Application

    OpenAIRE

    Tajima, Yoshiyuki; Onisawa, Takehisa

    2010-01-01

    In this chapter, the learning algorithm ME-FPRL is discussed. And it applied to the pursuit target task. Application results show that the ME-FPRL is more efficient than a RL or Modelbased RL. As a result, ME-FPRL is found to be able to apply to practical tasks. Our future work is constructing more an efficient learning system by using some advice and communication.

  11. Modeling and Simulation of Wide Area Communication for Centralized PMU-based Applications

    OpenAIRE

    Chenine, Moustafa; Nordström, Lars

    2011-01-01

    Phasor-based wide-area monitoring and control (WAMC) systems are becoming a reality with increased research, development, and deployments. Many potential control applications based on these systems are being proposed and researched. These applications are either local applications using data from one or a few phasor measurement units (PMUs) or centralized utilizing data from several PMUs. An aspect of these systems, which is less well researched, is the WAMC system's dependence on high-perfor...

  12. Applicability of an exposure model for the determination of emissions from mobile phone base stations

    International Nuclear Information System (INIS)

    Applicability of a model to estimate radiofrequency electromagnetic field (RF-EMF) strength in households from mobile phone base stations was evaluated with technical data of mobile phone base stations available from the German Net Agency, and dosimetric measurements, performed in an epidemiological study. Estimated exposure and exposure measured with dosemeters in 1322 participating households were compared. For that purpose, the upper 10. percentiles of both outcomes were defined as the 'higher exposed' groups. To assess the agreement of the defined 'higher' exposed groups, kappa coefficient, sensitivity and specificity were calculated. The present results show only a weak agreement of calculations and measurements (kappa values between -0.03 and 0.28, sensitivity between 7.1 and 34.6). Only in some of the sub-analyses, a higher agreement was found, e.g. when measured instead of interpolated geo-coordinates were used to calculate the distance between households and base stations, which is one important parameter in modelling exposure. During the development of the exposure model, more precise input data were available for its internal validation, which yielded kappa values between 0.41 and 0.68 and sensitivity between 55 and 76 for different types of housing areas. Contrary to this, the calculation of exposure - on the basis of the available imprecise data from the epidemiological study - is associated with a relatively high degree of uncertainty. Thus, the model can only be applied in epidemiological studies, when the uncertainty of the input data is considerably reduced. Otherwise, the use of dosemeters to determine the exposure from RF-EMF in epidemiological studies is recommended. (authors)

  13. A Simple Model for Estimating Evapotranspiration Based Solely on Remote Sensing: Algorithm and Application

    Science.gov (United States)

    Sun, Z.; Wang, Q.; Matsushita, B.; Fukushima, T.; Ouyang, Z.; Gebremichael, M.

    2009-12-01

    Remote sensing (RS) has been considered as the most promising tool for evapotranspiration (ET) estimations from local, regional to global scales. Many studies have been conducted to estimated ET using RS data, however, most of them are based partially on ground observations. This limits the applications of these algorithms when the necessary data are unavailable. Some other algorithms can generate real-time ET solely using remote sensing data, but lack mechanistic realism. In our study, we developed a new dual-source Simple Remote Sensing EvapoTranspiration model (Sim-ReSET) based only on RS data. One merit of this model is that the calculation of aerodynamic resistance can be avoided by means of a reference dry bare soil and an assumption that wind speed at the upper boundary of atmospheric surface layer is homogenous, but the aerodynamic characters are still considered by means of canopy height. The other merit is that all inputs (net radiation, soil heat flux, canopy height, variables related to land surface temperature) can be potentially obtained from remote sensing data, which allows obtaining regular RS-driven ET product. For the purposes of sensitivity analysis and performance evaluation of the Sim-ReSET model without the effect of potential uncertainties and errors from remote sensing data, the Sim-ReSET model was tested only using intensive ground observations at the Yucheng ecological station in the North China Plain from 2006 to 2008. Results show that the model has a good performance for instantaneous ET estimations with a mean absolute difference (MAD) of 34.27 W/m2 and a root mean square error (RMSE) of 41.84 W/m2 under neutral or near-neutral atmospheric conditions. On 12 cloudless days, the MAD of daily ET accumulated from instantaneous estimations is 0.26 mm/day, and the RMSE is 0.30 mm/day. In our study, we mapped Asian 16-day ET from 2000 to 2009 only using MODIS land data products based on the Sim-ReSET model. Then, the obtained ET product was

  14. Novel Component-Based Development Model for SIP-Based Mobile Application (1202)

    CERN Document Server

    Barnawi, Ahmed; Qureshi, M Rizwan Jameel; Khan, Asif Irshad

    2012-01-01

    Universities and Institutions these days' deals with issues related to with assessment of large number of students. Various evaluation methods have been adopted by examiners in different institutions to examining the ability of an individual, starting from manual means of using paper and pencil to electronic, from oral to written, practical to theoretical and many others. There is a need to expedite the process of examination in order to meet the increasing enrolment of students at the universities and institutes. Sip Based Mass Mobile Examination System (SiBMMES) expedites the examination process by automating various activities in an examination such as exam paper setting, Scheduling and allocating examination time and evaluation (auto-grading for objective questions) etc. SiBMMES uses the IP Multimedia Subsystem (IMS) that is an IP communications framework providing an environment for the rapid development of innovative and reusable services Session Initial Protocol (SIP) is a signalling (request-response)...

  15. Novel Component Based Development Model For Sip-Based Mobile Application

    CERN Document Server

    Barnawi, Ahmed; Qureshi, M Rizwan Jameel; Khan, Asif Irshad; 10.5121/ijsea.2012.3107

    2012-01-01

    Universities and Institutions these days' deals with issues related to with assessment of large number of students. Various evaluation methods have been adopted by examiners in different institutions to examining the ability of an individual, starting from manual means of using paper and pencil to electronic, from oral to written, practical to theoretical and many others. There is a need to expedite the process of examination in order to meet the increasing enrolment of students at the universities and institutes. Sip Based Mass Mobile Examination System (SiBMMES) expedites the examination process by automating various activities in an examination such as exam paper setting, Scheduling and allocating examination time and evaluation (auto-grading for objective questions) etc. SiBMMES uses the IP Multimedia Subsystem (IMS) that is an IP communications framework providing an environment for the rapid development of innovative and reusable services Session Initial Protocol (SIP) is a signalling (request-response)...

  16. Physiologically-based toxicokinetic modeling of zearalenone and its metabolites: application to the Jersey girl study.

    Directory of Open Access Journals (Sweden)

    Dwaipayan Mukherjee

    Full Text Available Zearalenone (ZEA, a fungal mycotoxin, and its metabolite zeranol (ZAL are known estrogen agonists in mammals, and are found as contaminants in food. Zeranol, which is more potent than ZEA and comparable in potency to estradiol, is also added as a growth additive in beef in the US and Canada. This article presents the development and application of a Physiologically-Based Toxicokinetic (PBTK model for ZEA and ZAL and their primary metabolites, zearalenol, zearalanone, and their conjugated glucuronides, for rats and for human subjects. The PBTK modeling study explicitly simulates critical metabolic pathways in the gastrointestinal and hepatic systems. Metabolic events such as dehydrogenation and glucuronidation of the chemicals, which have direct effects on the accumulation and elimination of the toxic compounds, have been quantified. The PBTK model considers urinary and fecal excretion and biliary recirculation and compares the predicted biomarkers of blood, urinary and fecal concentrations with published in vivo measurements in rats and human subjects. Additionally, the toxicokinetic model has been coupled with a novel probabilistic dietary exposure model and applied to the Jersey Girl Study (JGS, which involved measurement of mycoestrogens as urinary biomarkers, in a cohort of young girls in New Jersey, USA. A probabilistic exposure characterization for the study population has been conducted and the predicted urinary concentrations have been compared to measurements considering inter-individual physiological and dietary variability. The in vivo measurements from the JGS fall within the high and low predicted distributions of biomarker values corresponding to dietary exposure estimates calculated by the probabilistic modeling system. The work described here is the first of its kind to present a comprehensive framework developing estimates of potential exposures to mycotoxins and linking them with biologically relevant doses and biomarker

  17. Physiologically-Based Toxicokinetic Modeling of Zearalenone and Its Metabolites: Application to the Jersey Girl Study

    Science.gov (United States)

    Mukherjee, Dwaipayan; Royce, Steven G.; Alexander, Jocelyn A.; Buckley, Brian; Isukapalli, Sastry S.; Bandera, Elisa V.; Zarbl, Helmut; Georgopoulos, Panos G.

    2014-01-01

    Zearalenone (ZEA), a fungal mycotoxin, and its metabolite zeranol (ZAL) are known estrogen agonists in mammals, and are found as contaminants in food. Zeranol, which is more potent than ZEA and comparable in potency to estradiol, is also added as a growth additive in beef in the US and Canada. This article presents the development and application of a Physiologically-Based Toxicokinetic (PBTK) model for ZEA and ZAL and their primary metabolites, zearalenol, zearalanone, and their conjugated glucuronides, for rats and for human subjects. The PBTK modeling study explicitly simulates critical metabolic pathways in the gastrointestinal and hepatic systems. Metabolic events such as dehydrogenation and glucuronidation of the chemicals, which have direct effects on the accumulation and elimination of the toxic compounds, have been quantified. The PBTK model considers urinary and fecal excretion and biliary recirculation and compares the predicted biomarkers of blood, urinary and fecal concentrations with published in vivo measurements in rats and human subjects. Additionally, the toxicokinetic model has been coupled with a novel probabilistic dietary exposure model and applied to the Jersey Girl Study (JGS), which involved measurement of mycoestrogens as urinary biomarkers, in a cohort of young girls in New Jersey, USA. A probabilistic exposure characterization for the study population has been conducted and the predicted urinary concentrations have been compared to measurements considering inter-individual physiological and dietary variability. The in vivo measurements from the JGS fall within the high and low predicted distributions of biomarker values corresponding to dietary exposure estimates calculated by the probabilistic modeling system. The work described here is the first of its kind to present a comprehensive framework developing estimates of potential exposures to mycotoxins and linking them with biologically relevant doses and biomarker measurements

  18. Modeling dependence based on mixture copulas and its application in risk management

    Institute of Scientific and Technical Information of China (English)

    OUYANG Zi-sheng; LIAO Hui; YANG Xiang-qun

    2009-01-01

    This paper is concerned with the statistical modeling of the dependence structure of multivariate financial data using the copula, and the application of copula functions in VaR valuation. After the introduction of the pure copula method and the maximum and minimum mixture copula method, authors present a new algorithm based on the more generalized mixture copula functions and the dependence measure, and apply the method to the portfolio of Shanghai stock composite index and Shenzhen stock component index. Comparing with the results from various methods, one can find that the mixture copula method is better than the pure Gaussia copula method and the maximum and minimum mixture copula method on different VaR level.

  19. Genetic Modeling of GIS-Based Cell Clusters and Its Application in Mineral Resources Prediction

    Institute of Scientific and Technical Information of China (English)

    2003-01-01

    This paper presents a synthetic analysis method for multi-sourced geological data from geographic information system (GIS). In the previous practices of mineral resources prediction, a usually adopted methodology has been statistical analysis of cells delimitated based on thoughts of random sampiing. That might lead to insufficient utilization of local spatial information, for a cell is treated as a point without internal structure. We now take "cell clusters", i. e. , spatial associations of cells, as basic units of statistics, thus the spatial configuration information of geological variables is easier to be detected and utilized, and the accuracy and reliability of prediction are improved. We build a linear multi-discriminating model for the clusters via genetic algorithm. Both the right-judgment rates and the in-class vs. between-class distance ratios are considered to form the evolutional adaptive values of the population. An application of the method in gold mineral resources prediction in east Xinjiang, China is presented.

  20. Development, fabrication, and modeling of highly sensitive conjugated polymer based piezoresistive sensors in electronic skin applications

    Science.gov (United States)

    Khalili, Nazanin; Naguib, Hani E.; Kwon, Roy H.

    2016-04-01

    Human intervention can be replaced through development of tools resulted from utilizing sensing devices possessing a wide range of applications including humanoid robots or remote and minimally invasive surgeries. Similar to the five human senses, sensors interface with their surroundings to stimulate a suitable response or action. The sense of touch which arises in human skin is among the most challenging senses to emulate due to its ultra high sensitivity. This has brought forth novel challenging issues to consider in the field of biomimetic robotics. In this work, using a multiphase reaction, a polypyrrole (PPy) based hydrogel is developed as a resistive type pressure sensor with an intrinsically elastic microstructure stemming from three dimensional hollow spheres. Furthermore, a semi-analytical constriction resistance model accounting for the real contact area between the PPy hydrogel sensors and the electrode along with the dependency of the contact resistance change on the applied load is developed. The model is then solved using a Monte Carlo technique and the sensitivity of the sensor is obtained. The experimental results showed the good tracking ability of the proposed model.

  1. Agent-based land-use models: a review of applications

    OpenAIRE

    Matthews, RB; Gilbert, NG; Roach, A.; Polhill, JG; Gotts, NM

    2007-01-01

    Agent-based modelling is an approach that has been receiving attention by the land use modelling community in recent years, mainly because it offers a way of incorporating the influence of human decision-making on land use in a mechanistic, formal, and spatially explicit way, taking into account social interaction, adaptation, and decision-making at different levels. Specific advantages of agent-based models include their ability to model individual decision-making entities and their interact...

  2. Central Puget Sound Ecopath/Ecosim model biological parameters - Developing food web models for ecosystem-based management applications in Puget Sound

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This project is developing food web models for ecosystem-based management applications in Puget Sound. It is primarily being done by NMFS FTEs and contractors, in...

  3. Research on urban rapid 3D modeling and application based on CGA rule

    Science.gov (United States)

    Li, Jing-wen; Jiang, Jian-wu; Zhou, Song; Yin, Shou-qiang

    2015-12-01

    Use CityEngine as the 3D modeling platform, research on urban rapid 3D modeling technology based on the CGA(Computer Generated Architectur) rule , solved the problem of the rapid creation of urban 3D model in large scenes , and research on building texture processing and 3D model optimization techniques based on CGA rule , using component modeling method , solved the problem of texture distortion and model redundancy in the traditional fast modeling 3D model , and development of a three-dimensional view and analysis system based on ArcGIS Engine , realization of 3D model query , distance measurement , specific path flight , 3D marking , Scene export,etc.

  4. Application of GIS based data driven evidential belief function model to predict groundwater potential zonation

    Science.gov (United States)

    Nampak, Haleh; Pradhan, Biswajeet; Manap, Mohammad Abd

    2014-05-01

    The objective of this paper is to exploit potential application of an evidential belief function (EBF) model for spatial prediction of groundwater productivity at Langat basin area, Malaysia using geographic information system (GIS) technique. About 125 groundwater yield data were collected from well locations. Subsequently, the groundwater yield was divided into high (⩾11 m3/h) and low yields (validation purpose. To perform cross validation, the frequency ratio (FR) approach was applied into remaining groundwater wells with low yield to show the spatial correlation between the low potential zones of groundwater productivity. A total of twelve groundwater conditioning factors that affect the storage of groundwater occurrences were derived from various data sources such as satellite based imagery, topographic maps and associated database. Those twelve groundwater conditioning factors are elevation, slope, curvature, stream power index (SPI), topographic wetness index (TWI), drainage density, lithology, lineament density, land use, normalized difference vegetation index (NDVI), soil and rainfall. Subsequently, the Dempster-Shafer theory of evidence model was applied to prepare the groundwater potential map. Finally, the result of groundwater potential map derived from belief map was validated using testing data. Furthermore, to compare the performance of the EBF result, logistic regression model was applied. The success-rate and prediction-rate curves were computed to estimate the efficiency of the employed EBF model compared to LR method. The validation results demonstrated that the success-rate for EBF and LR methods were 83% and 82% respectively. The area under the curve for prediction-rate of EBF and LR methods were calculated 78% and 72% respectively. The outputs achieved from the current research proved the efficiency of EBF in groundwater potential mapping.

  5. SAR Imagery Simulation of Ship Based on Electromagnetic Calculations and Sea Clutter Modelling for Classification Applications

    International Nuclear Information System (INIS)

    Ship detection and classification with space-borne SAR has many potential applications within the maritime surveillance, fishery activity management, monitoring ship traffic, and military security. While ship detection techniques with SAR imagery are well established, ship classification is still an open issue. One of the main reasons may be ascribed to the difficulties on acquiring the required quantities of real data of vessels under different observation and environmental conditions with precise ground truth. Therefore, simulation of SAR images with high scenario flexibility and reasonable computation costs is compulsory for ship classification algorithms development. However, the simulation of SAR imagery of ship over sea surface is challenging. Though great efforts have been devoted to tackle this difficult problem, it is far from being conquered. This paper proposes a novel scheme for SAR imagery simulation of ship over sea surface. The simulation is implemented based on high frequency electromagnetic calculations methods of PO, MEC, PTD and GO. SAR imagery of sea clutter is modelled by the representative K-distribution clutter model. Then, the simulated SAR imagery of ship can be produced by inserting the simulated SAR imagery chips of ship into the SAR imagery of sea clutter. The proposed scheme has been validated with canonical and complex ship targets over a typical sea scene

  6. Enhancement Factors in Ozone Absorption Based on the Surface Renewal Model and its Application

    Institute of Scientific and Technical Information of China (English)

    2000-01-01

    Based on the Danckwerts surface renewal model, a simple explicit expression of theenhancement factor in ozone absorption with a first order ozone self-decomposition and parallel secondorder ozonation reactions has been derived. The results are compared with our previous work based onthe film theory. The 2,4-dichlorophenol destruction rate by ozonation is predicted using the enhancementfactor model in this paper.

  7. Predictive models based on sensitivity theory and their application to practical shielding problems

    International Nuclear Information System (INIS)

    Two new calculational models based on the use of cross-section sensitivity coefficients have been devised for calculating radiation transport in relatively simple shields. The two models, one an exponential model and the other a power model, have been applied, together with the traditional linear model, to 1- and 2-m-thick concrete-slab problems in which the water content, reinforcing-steel content, or composition of the concrete was varied. Comparing the results obtained with the three models with those obtained from exact one-dimensional discrete-ordinates transport calculations indicates that the exponential model, named the BEST model (for basic exponential shielding trend), is a particularly promising predictive tool for shielding problems dominated by exponential attenuation. When applied to a deep-penetration sodium problem, the BEST model also yields better results than do calculations based on second-order sensitivity theory

  8. Gradient-based Kriging approximate model and its application research to optimization design

    Institute of Scientific and Technical Information of China (English)

    XUAN Ying; XIANG JunHua; ZHANG WeiHua; ZHANG YuLin

    2009-01-01

    In the process of multidisciplinary design optimization, there exits the calculation complexity problem due to frequently calling high fidelity system analysis models. The high fidelity system analysis models can be surrogated by approximate models. The sensitivity analysis and numerical noise filtering can be done easily by coupling approximate models to optimization. Approximate models can reduce the number of executions of the problem's simulation code during optimization, so the solution efficiency of the multidisciplinary design optimization problem can be improved. Most optimization methods are based on gradient. The gradients of the objective and constrain functions are gained easily. The gradient-based Kriging (GBK) approximate model can be constructed by using system response value and its gradients. The gradients can greatly improve prediction precision of system response. The hybrid optimization method is constructed by coupling GBK approximate models to gradient-based optimization methods. An aircraft aerodynamics shape optimization design example indicates that the methods of this paper can achieve good feasibility and validity.

  9. A Model for Protein Sequence Evolution Based on Selective Pressure for Protein Stability: Application to Hemoglobins

    OpenAIRE

    Lorraine Marsh

    2009-01-01

    Negative selection against protein instability is a central influence on evolution of proteins. Protein stability is maintained over evolution despite changes in underlying sequences. An empirical all-site stability-based model of evolution was developed to focus on the selection of residues arising from their contributions to protein stability. In this model, site rates could vary. A structure-based method was used to predict stationary frequencies of hemoglobin residues based on their prope...

  10. Security Model for Microsoft Based Mobile Sales Management Application in Private Cloud Computing

    OpenAIRE

    Kuan Chee Houng; Bharanidharan Shanmugam; Ganthan Narayana Samy; Sameer Hasan Albakri; Azuan Ahmad

    2013-01-01

    The Microsoft-based mobile sales management application is a sales force management application that currently running on Windows Mobile 6.5. It handles sales-related activity and cuts down the administrative task of sales representative. Then, Windows launch a new mobile operating system, Windows Phone and stop providing support to Windows Mobile. This has become an obstacle for Windows Mobile development. From time to time, Windows Mobile will be eliminated from the market due to no support...

  11. Brain Emotional Learning-Based Prediction Model (For Long-Term Chaotic Prediction Applications)

    OpenAIRE

    Parsapoor, Mahboobeh

    2016-01-01

    This study suggests a new prediction model for chaotic time series inspired by the brain emotional learning of mammals. We describe the structure and function of this model, which is referred to as BELPM (Brain Emotional Learning-Based Prediction Model). Structurally, the model mimics the connection between the regions of the limbic system, and functionally it uses weighted k nearest neighbors to imitate the roles of those regions. The learning algorithm of BELPM is defined using steepest des...

  12. Proceedings First Workshop on Applications of Membrane computing, Concurrency and Agent-based modelling in POPulation biology

    CERN Document Server

    Milazzo, Paolo; 10.4204/EPTCS.33

    2010-01-01

    This volume contains the papers presented at the first International Workshop on Applications of Membrane Computing, Concurrency and Agent-based Modelling in Population Biology (AMCA-POP 2010) held in Jena, Germany on August 25th, 2010 as a satellite event of the 11th Conference on Membrane Computing (CMC11). The aim of the workshop is to investigate whether formal modelling and analysis techniques could be applied with profit to systems of interest for population biology and ecology. The considered modelling notations include membrane systems, Petri nets, agent-based notations, process calculi, automata-based notations, rewriting systems and cellular automata. Such notations enable the application of analysis techniques such as simulation, model checking, abstract interpretation and type systems to study systems of interest in disciplines such as population biology, ecosystem science, epidemiology, genetics, sustainability science, evolution and other disciplines in which population dynamics and interactions...

  13. Inverse Modeling of Human Knee Joint Based on Geometry and Vision Systems for Exoskeleton Applications

    Directory of Open Access Journals (Sweden)

    Eduardo Piña-Martínez

    2015-01-01

    Full Text Available Current trends in Robotics aim to close the gap that separates technology and humans, bringing novel robotic devices in order to improve human performance. Although robotic exoskeletons represent a breakthrough in mobility enhancement, there are design challenges related to the forces exerted to the users’ joints that result in severe injuries. This occurs due to the fact that most of the current developments consider the joints as noninvariant rotational axes. This paper proposes the use of commercial vision systems in order to perform biomimetic joint design for robotic exoskeletons. This work proposes a kinematic model based on irregular shaped cams as the joint mechanism that emulates the bone-to-bone joints in the human body. The paper follows a geometric approach for determining the location of the instantaneous center of rotation in order to design the cam contours. Furthermore, the use of a commercial vision system is proposed as the main measurement tool due to its noninvasive feature and for allowing subjects under measurement to move freely. The application of this method resulted in relevant information about the displacements of the instantaneous center of rotation at the human knee joint.

  14. Convergence rates for rank-based models with applications to portfolio theory

    CERN Document Server

    Ichiba, Tomoyuki; Shkolnikov, Mykhaylo

    2011-01-01

    We determine rates of convergence of rank-based interacting diffusions and semimartingale reflecting Brownian motions to equilibrium. Convergence rate for the total variation metric is derived using Lyapunov functions. Sharp fluctuations of additive functionals are obtained using Transportation Cost-Information inequalities for Markov processes. We work out various applications to the rank-based abstract equity markets used in Stochastic Portfolio Theory. For example, we produce quantitative bounds, including constants, for fluctuations of market weights and occupation times of various ranks for individual coordinates. Another important application is the comparison of performance between symmetric functionally generated portfolios and the market portfolio. This produces estimates of probabilities of "beating the market".

  15. Tensor Product Model Transformation-based Controller Design for Gantry Crane Control System – An Application Approach

    Directory of Open Access Journals (Sweden)

    Fetah Kolonic

    2006-10-01

    Full Text Available The Tensor Product (TP model transformation is a recently proposed techniquefor transforming given Linear Parameter Varying (LPV state-space models into polytopicmodel form, namely, to parameter varying convex combination of Linear Time Invariant(LTI systems. The main advantage of the TP model transformation is that it is executablein a few minutes and the Linear Matrix Inequality (LMI-based control design frameworkscan immediately be applied to the resulting polytopc models to yield controllers withtractable and guaranteed performance. Various applications of the TP modeltransformation-based design were studied via academic complex and benchmark problems,but no real experimental environment-based study was published. Thus, the main objectiveof this paper is to study how the TP model transformation performs in a real world problemand control setup. The laboratory concept for TP model-based controller design,simulation and real time running on an electromechanical system is presented.Development system for TP model-based controller with one hardware/software platformand target system with real-time hardware/ software support are connected in the uniquesystem. Proposed system is based on microprocessor of personal computer (PC forsimulation and software development as well as for real-time control. Control algorithm,designed and simulated in MATLAB/SIMULINK environment, use graphically orientedsoftware interface for real-time code generation. Some specific conflicting industrial tasksin real industrial crane application, such as fast load positioning control and load swingangle minimization, are considered and compared with other controller types.

  16. Fish-eye Camera Calibration Model Based on Vector Observations and Its Application

    Directory of Open Access Journals (Sweden)

    ZHAN Yinhu

    2016-03-01

    Full Text Available A fish-eye camera calibration model is presented, basic observations of which consist of both half angle of view and azimuth. Rodrigues matrix is introduced into the model, and three Rodrigues parameters instead of Euler angles are used to represent elements of exterior orientation in order to simplify the expressions and calculations of observation equations.The new model is compared with the existing models based on half angle of view constraint by actual star-map data processing, and the results indicate that the model is superior to control the azimuth error, while slightly inferior to constrain the error of half angle of view. It is advised that radial distortion parameters should be determined by the model based on half angle of view constraint at first, and other camera parameters should be calculated by the new model.

  17. Development and application of a model for analysis and design phases of Web-based system development

    Institute of Scientific and Technical Information of China (English)

    2003-01-01

    Despite a short history of the Web development, Web-related technologies are rapidly develop-ing. However, the Web application quality is improving slowly, which requires efficient methods for devel-oping Web systems. This study presents a model for Web-based software development for analysis and design phases based on the ISO/IEC 12207 standard. It describes the methods used to define processes and entities in order to reflect the contents in Web applications. It applies the methodology of Web-Road Map by KCC Information and Technology using this model to the public project. As a result, Web-Road Map is proven to be an efficient model to analyze and design Web-applications.

  18. Methodology and applications in non-linear model-based geostatistics

    DEFF Research Database (Denmark)

    Christensen, Ole Fredslund

    Today geostatistics is used in a number of research areas, among others agricultural and environmental sciences.This thesis concerns data and applications where the classical Gaussian spatial model is not appropriate. A transformation could be used in an attempt to obtain data that are approximat...

  19. Dynamic Phasors-Based Modeling and Stability Analysis of Droop-Controlled Inverters for Microgrid Applications

    DEFF Research Database (Denmark)

    Guo, Xiaoqiang; Lu, Zhigang; Wang, Baocheng; Sun, Xiaofeng; Wang, Lei; Guerrero, Josep M.

    2014-01-01

    System modeling and stability analysis is one of the most important issues of inverter-dominated microgrids. It is useful to determine the system stability and optimize the control parameters. The complete small signal models for the inverter-dominated microgrids have been developed which are very...... accurate and could be found in literature. However, the modeling procedure will become very complex when the number of inverters in microgrid is large. One possible solution is to use the reduced-order small signal models for the inverter-dominated microgrids. Unfortunately, the reduced-order small signal...... of the system, while the conventional reduced-order small signal model fails. In addition, the virtual ω-E frame power control method, which deals with the power coupling caused by the line impedance X/R characteristic, has also been chosen as an application example of the proposed modeling technique....

  20. Reliability Modeling Development and Its Applications for Ceramic Capacitors with Base-Metal Electrodes (BMEs)

    Science.gov (United States)

    Liu, Donhang

    2014-01-01

    This presentation includes a summary of NEPP-funded deliverables for the Base-Metal Electrodes (BMEs) capacitor task, development of a general reliability model for BME capacitors, and a summary and future work.

  1. Energy Minimization-Based Spatially Constrained Mixture Model and its Application to Image Segmentation

    Directory of Open Access Journals (Sweden)

    Zhiyong Xiao

    2015-08-01

    Full Text Available A novel method is proposed for image segmentation based on probabilistic field theory. This model assumes that the whole pixels of an image and some unknown parameters form a field. According to this model, the pixel labels are generated by a compound function of the field. The main novelty of this model is it consider the features of the pixels and the interdependent among the pixels. The parameters are generated by a novel spatially variant mixture model and estimated by expectation-maximization (EM- based algorithm. Thus, we simultaneously impose the spatial smoothness on the prior knowledge. Numerical experiments are presented where the proposed method and other mixture model-based methods were tested on synthetic and real world images. These experimental results demonstrate that our algorithm achieves competitive performance compared to other methods.

  2. Development and application of compact models of packages based on DELPHI methodology

    CERN Document Server

    Parry, J; Shidore, S

    1997-01-01

    The accurate prediction of the temperatures of critical electronic parts at the package- board- and system-level is seriously hampered by the lack of reliable, standardised input data for the characterisation of the thermal $9 behaviour of these parts. The recently completed collaborative European project, DELPHI has been concerned with the creation and experimental validation of thermal models (both detailed and compact) of a range of electronic parts, $9 including mono-chip packages. This paper demonstrates the reliable performance of thermal compact models in a range of applications, by comparison with the detailed models from which they were derived. (31 refs).

  3. Enhancement Factors in Ozone Absorption Based on the Surface Renewal Model and its Application

    Institute of Scientific and Technical Information of China (English)

    程江; 杨卓如; 陈焕钦; C.H.Kuo; M.E.Zappi

    2000-01-01

    Based on the Danckwerts surface renewal model, a simple explicit expression of the enhancement factor in ozone absorption with a first order ozone self-decomposition and parallel second order ozonation reactions has been derived. The results are compared with our previous work based on the film theory. The 2,4-dichlorophenol destruction rate by ozonation is predicted using the enhancement factor model in this paper.

  4. Finite mathematics models and applications

    CERN Document Server

    Morris, Carla C

    2015-01-01

    Features step-by-step examples based on actual data and connects fundamental mathematical modeling skills and decision making concepts to everyday applicability Featuring key linear programming, matrix, and probability concepts, Finite Mathematics: Models and Applications emphasizes cross-disciplinary applications that relate mathematics to everyday life. The book provides a unique combination of practical mathematical applications to illustrate the wide use of mathematics in fields ranging from business, economics, finance, management, operations research, and the life and social sciences.

  5. An Empirically Based Method of Q-Matrix Validation for the DINA Model: Development and Applications

    Science.gov (United States)

    de la Torre, Jimmy

    2008-01-01

    Most model fit analyses in cognitive diagnosis assume that a Q matrix is correct after it has been constructed, without verifying its appropriateness. Consequently, any model misfit attributable to the Q matrix cannot be addressed and remedied. To address this concern, this paper proposes an empirically based method of validating a Q matrix used…

  6. A new tool to control meat products safety: a web based application of predictive microbiology models

    OpenAIRE

    Delhalle, Laurent; Adolphe, Ysabelle; Crevecoeur, Sébastien; Didimo Imazaki, Pedro Henrique; Daube, Georges; Clinquart, Antoine

    2011-01-01

    Predictive microbiology is considered by the European legislation as a tool to control food safety. Meat and meat products are particularly sensitive to contamination with pathogens. However, development of predictive microbiology models and interpretation of results require specific knowledge. A free web based model has been developed for an easy use by people who are not experts in this field as industries and public authorities. The model can simulate the growth of Salmonella spp, Listeria...

  7. Application of a hazard-based visual predictive check to evaluate parametric hazard models.

    Science.gov (United States)

    Huh, Yeamin; Hutmacher, Matthew M

    2016-02-01

    Parametric models used in time to event analyses are evaluated typically by survival-based visual predictive checks (VPC). Kaplan-Meier survival curves for the observed data are compared with those estimated using model-simulated data. Because the derivative of the log of the survival curve is related to the hazard--the typical quantity modeled in parametric analysis--isolation, interpretation and correction of deficiencies in the hazard model determined by inspection of survival-based VPC's is indirect and thus more difficult. The purpose of this study is to assess the performance of nonparametric hazard estimators of hazard functions to evaluate their viability as VPC diagnostics. Histogram-based and kernel-smoothing estimators were evaluated in terms of bias of estimating the hazard for Weibull and bathtub-shape hazard scenarios. After the evaluation of bias, these nonparametric estimators were assessed as a method for VPC evaluation of the hazard model. The results showed that nonparametric hazard estimators performed reasonably at the sample sizes studied with greater bias near the boundaries (time equal to 0 and last observation) as expected. Flexible bandwidth and boundary correction methods reduced these biases. All the nonparametric estimators indicated a misfit of the Weibull model when the true hazard was a bathtub shape. Overall, hazard-based VPC plots enabled more direct interpretation of the VPC results compared to survival-based VPC plots. PMID:26563504

  8. Advancement of Satellite-based Rainfall Applications for Hydrologic Modeling in Topographically Complex Regions

    Science.gov (United States)

    Yilmaz, Koray; Derin, Yagmur

    2014-05-01

    Accuracy and reliability of hydrological modeling studies heavily depends on quality and availability of precipitation estimates. However hydrological studies in developing countries, especially over complex topography, are limited due to unavailability and scarcity of ground-based networks. In this study we evaluate three different satellite-based rainfall retrieval algorithms namely, Tropical Rainfall Measuring Mission Multi-satellite Precipitation Analysis (TMPA), NOAA/Climate Prediction Center Morphing Method (CMORPH) and EUMETSAT's Multi-Sensor Precipitation Estimate (MPE) over orographically complex Western Black Sea Basin in Turkey, using a relatively dense rain gauge network. Our results indicated that satellite-based products significantly underestimated the rainfall in regions characterized by orographic rainfall and overestimated the rainfall in the drier regions with seasonal dependency. Further, we devised a new bias adjustment algorithm for the satellite-based rainfall products based on the "physiographic similarity" concept. Our results showed that proposed bias adjustment algorithm is better suited to regions with complex topography and provided improved results compared to the baseline "inverse distance weighting" method. To evaluate the utility of satellite-based products in hydrologic modeling studies, we implemented the MIKE SHE-MIKE 11 integrated fully distributed physically based hydrological model in the study region driven by ground-based and satellite-based precipitation estimates. Model parameter estimation was performed using a constrained calibration approach guided by multiple "signature measures" to estimate model parameters in a hydrologically meaningful way rather than using the traditional "statistical" objective functions that largely mask valuable hydrologic information during calibration process. In this presentation we will provide a discussion of evaluation and bias correction of the satellite-based precipitation products and

  9. Advances In Global Aerosol Modeling Applications Through Assimilation of Satellite-Based Lidar Measurements

    Science.gov (United States)

    Campbell, James; Hyer, Edward; Zhang, Jianglong; Reid, Jeffrey; Westphal, Douglas; Xian, Peng; Vaughan, Mark

    2010-05-01

    Modeling the instantaneous three-dimensional aerosol field and its downwind transport represents an endeavor with many practical benefits foreseeable to air quality, aviation, military and science agencies. The recent proliferation of multi-spectral active and passive satellite-based instruments measuring aerosol physical properties has served as an opportunity to develop and refine the techniques necessary to make such numerical modeling applications possible. Spurred by high-resolution global mapping of aerosol source regions, and combined with novel multivariate data assimilation techniques designed to consider these new data streams, operational forecasts of visibility and aerosol optical depths are now available in near real-time1. Active satellite-based aerosol profiling, accomplished using lidar instruments, represents a critical element for accurate analysis and transport modeling. Aerosol source functions, alone, can be limited in representing the macrophysical structure of injection scenarios within a model. Two-dimensional variational (2D-VAR; x, y) assimilation of aerosol optical depth from passive satellite observations significantly improves the analysis of the initial state. However, this procedure can not fully compensate for any potential vertical redistribution of mass required at the innovation step. The expense of an inaccurate vertical analysis of aerosol structure is corresponding errors downwind, since trajectory paths within successive forecast runs will likely diverge with height. In this paper, the application of a newly-designed system for 3D-VAR (x,y,z) assimilation of vertical aerosol extinction profiles derived from elastic-scattering lidar measurements is described [Campbell et al., 2009]. Performance is evaluated for use with the U. S. Navy Aerosol Analysis and Prediction System (NAAPS) by assimilating NASA/CNES satellite-borne Cloud-Aerosol Lidar with Orthogonal Polarization (CALIOP) 0.532 μm measurements [Winker et al., 2009

  10. Variable Selection and Updating In Model-Based Discriminant Analysis for High Dimensional Data with Food Authenticity Applications*

    OpenAIRE

    Murphy, Thomas Brendan; Dean, Nema; Raftery, Adrian E.

    2010-01-01

    Food authenticity studies are concerned with determining if food samples have been correctly labeled or not. Discriminant analysis methods are an integral part of the methodology for food authentication. Motivated by food authenticity applications, a model-based discriminant analysis method that includes variable selection is presented. The discriminant analysis model is fitted in a semi-supervised manner using both labeled and unlabeled data. The method is shown to give ...

  11. A Danger Theory Based Mobile Virus Detection Model and Its Application in Inhibiting Virus

    Directory of Open Access Journals (Sweden)

    Tianliang Lu

    2012-08-01

    Full Text Available According to the propagation and destruction characteristics of mobile phone viruses, a virus detection model based on the Danger Theory is proposed. This model includes four phases: danger capture, antigen presentation, antibody generation and antibody distribution. In this model, local knowledge of mobile phones is exploited by the agents that are running in mobile phones to discover danger caused by viruses. The Antigen Presenting Cells (APCs present the antigen from mobile phones in the danger zone, and the Decision Center confirms the infection of viruses. After the antibody is generated by self-tolerating using the negative selection algorithm, the Decision Center distributes the antibody to mobile phones. Due to the distributed and cooperative mechanism of artificial immune system, the proposed model lowers the storage and computing consumption of mobile phones. The simulation results show that based on the mobile phone virus detection model, the proposed virus immunization strategy can effectively inhibit the propagation of mobile phone viruses.

  12. FANN-Based Surface Water Quality Evaluation Model and Its Application in the Shaoguan Area

    Institute of Scientific and Technical Information of China (English)

    YANG Meini; LI Dingfang; YANG Jinbo; XIONG Wei

    2007-01-01

    A fuzzy neural network model is proposed to evaluate water quality. The model contains two parts: first, fuzzy mathematics theory is used to standardize the samples; second, the RBF neural network and the BP neural network are used to train the standardized samples. The proposed model was applied to assess the water quality of 16 sections in 9 rivers in the Shaoguan area in 2005. The evaluation result was compared with that of the RBF neural network method and the reported results in the Shaoguan area in 2005. It indicated that the performance of the proposed fuzzy neural network model is practically feasible in the application of water quality assessment and its operation is simple.

  13. Scenario-based, closed-loop model predictive control with application to emergency vehicle scheduling

    Science.gov (United States)

    Goodwin, Graham. C.; Medioli, Adrian. M.

    2013-08-01

    Model predictive control has been a major success story in process control. More recently, the methodology has been used in other contexts, including automotive engine control, power electronics and telecommunications. Most applications focus on set-point tracking and use single-sequence optimisation. Here we consider an alternative class of problems motivated by the scheduling of emergency vehicles. Here disturbances are the dominant feature. We develop a novel closed-loop model predictive control strategy aimed at this class of problems. We motivate, and illustrate, the ideas via the problem of fluid deployment of ambulance resources.

  14. Modeling and Deployment of Model-Based Decentralized Embedded Diagnosis inside Vehicles: Application to Smart Distance Keeping Function

    Directory of Open Access Journals (Sweden)

    Othman Nasri

    2012-01-01

    Full Text Available The deployment of a fault diagnosis strategy in the Smart Distance Keeping (SDK system with a decentralized architecture is presented. The SDK system is an advanced Adaptive Cruise Control (ACC system implemented in a Renault-Volvo Trucks vehicle to increase safety by overcoming some ACC limitations. One of the main differences between this new system and the classical ACC is the choice of the safe distance. This latter is the distance between the vehicle equipped with the ACC or the SDK system and the obstacle-in-front (which may be another vehicle. It is supposed fixed in the case of the ACC, while variable in the case of the SDK. The variation of this distance depends essentially on the relative velocity between the vehicle and the obstacle-in-front. The main goal of this work is to analyze measurements, issued from the SDK elements, in order to detect, to localize, and to identify some faults that may occur. Our main contribution is the proposition of a decentralized approach permitting to carry out an on-line diagnosis without computing the global model and to achieve most of the work locally avoiding huge extra diagnostic information traffic between components. After a detailed description of the SDK system, this paper explains the model-based decentralized solution and its application to the embedded diagnosis of the SDK system inside Renault-Volvo Truck with five control units connected via a CAN-bus using “Hardware in the Loop” (HIL technique. We also discuss the constraints that must be fulfilled.

  15. Multi-scale Modelling of bcc-Fe Based Alloys for Nuclear Applications

    International Nuclear Information System (INIS)

    , advanced techniques to fit interatomic potentials consistent with thermodynamics are proposed and the results of their application to the mentioned alloys are presented. Next, the development of advanced methods, based on the use of artificial intelligence, to improve both the physical reliability and the computational efficiency of kinetic Monte Carlo codes for the study of point-defect clustering and phase changes beyond the scale of MD, is reported. These recent progresses bear the promise of being able, in the near future, of producing reliable tools for the description of the microstructure evolution of realistic model alloys under irradiation. (author)

  16. Object-oriented modelling with unified modelling language 2.0 for simple software application based on agile methodology

    CERN Document Server

    Warnars, Spits

    2010-01-01

    Unified modelling language (UML) 2.0 introduced in 2002 has been developing and influencing object-oriented software engineering and has become a standard and reference for information system analysis and design modelling. There are many concepts and theories to model the information system or software application with UML 2.0, which can make ambiguities and inconsistencies for a novice to learn to how to model the system with UML especially with UML 2.0. This article will discuss how to model the simple software application by using some of the diagrams of UML 2.0 and not by using the whole diagrams as suggested by agile methodology. Agile methodology is considered as convenient for novices because it can deliver the information technology environment to the end-user quickly and adaptively with minimal documentation. It also has the ability to deliver best performance software application according to the customer's needs. Agile methodology will make simple model with simple documentation, simple team and si...

  17. Gradient-based Kriging approximate model and its application research to optimization design

    Institute of Scientific and Technical Information of China (English)

    2009-01-01

    In the process of multidisciplinary design optimization, there exits the calculation complexity problem due to frequently calling high fidelity system analysis models. The high fidelity system analysis models can be surrogated by approximate models. The sensitivity analysis and numerical noise filtering can be done easily by coupling approximate models to optimization. Approximate models can reduce the number of executions of the problem’s simulation code during optimization, so the solution efficiency of the multidisciplinary design optimization problem can be improved. Most optimization methods are based on gradient. The gradients of the objective and constrain functions are gained easily. The gra- dient-based Kriging (GBK) approximate model can be constructed by using system response value and its gradients. The gradients can greatly improve prediction precision of system response. The hybrid optimization method is constructed by coupling GBK approximate models to gradient-based optimiza- tion methods. An aircraft aerodynamics shape optimization design example indicates that the methods of this paper can achieve good feasibility and validity.

  18. Developments in model-based optimization and control distributed control and industrial applications

    CERN Document Server

    Grancharova, Alexandra; Pereira, Fernando

    2015-01-01

    This book deals with optimization methods as tools for decision making and control in the presence of model uncertainty. It is oriented to the use of these tools in engineering, specifically in automatic control design with all its components: analysis of dynamical systems, identification problems, and feedback control design. Developments in Model-Based Optimization and Control takes advantage of optimization-based formulations for such classical feedback design objectives as stability, performance and feasibility, afforded by the established body of results and methodologies constituting optimal control theory. It makes particular use of the popular formulation known as predictive control or receding-horizon optimization. The individual contributions in this volume are wide-ranging in subject matter but coordinated within a five-part structure covering material on: · complexity and structure in model predictive control (MPC); · collaborative MPC; · distributed MPC; · optimization-based analysis and desi...

  19. Application of Transfer Matrix Approach to Modeling and Decentralized Control of Lattice-Based Structures

    Science.gov (United States)

    Cramer, Nick; Swei, Sean Shan-Min; Cheung, Kenny; Teodorescu, Mircea

    2015-01-01

    This paper presents a modeling and control of aerostructure developed by lattice-based cellular materials/components. The proposed aerostructure concept leverages a building block strategy for lattice-based components which provide great adaptability to varying ight scenarios, the needs of which are essential for in- ight wing shaping control. A decentralized structural control design is proposed that utilizes discrete-time lumped mass transfer matrix method (DT-LM-TMM). The objective is to develop an e ective reduced order model through DT-LM-TMM that can be used to design a decentralized controller for the structural control of a wing. The proposed approach developed in this paper shows that, as far as the performance of overall structural system is concerned, the reduced order model can be as e ective as the full order model in designing an optimal stabilizing controller.

  20. Nationwide impact and vehicle to grid application of electric vehicles mobility using an activity based model

    OpenAIRE

    Álvaro, Roberto; González, Jairo; Fraile Ardanuy, José Jesús; Knapen, Luk; JANSSENS, Davy

    2013-01-01

    This paper describes the impact of electric mobility on the transmission grid in Flanders region (Belgium), using a micro-simulation activity based models. These models are used to provide temporal and spatial estimation of energy and power demanded by electric vehicles (EVs) in different mobility zones. The increment in the load demand due to electric mobility is added to the background load demand in these mobility areas and the effects over the transmission substations are analyzed. From t...

  1. Spatial Rule-Based Modeling: A Method and Its Application to the Human Mitotic Kinetochore

    Directory of Open Access Journals (Sweden)

    Jan Huwald

    2013-07-01

    Full Text Available A common problem in the analysis of biological systems is the combinatorial explosion that emerges from the complexity of multi-protein assemblies. Conventional formalisms, like differential equations, Boolean networks and Bayesian networks, are unsuitable for dealing with the combinatorial explosion, because they are designed for a restricted state space with fixed dimensionality. To overcome this problem, the rule-based modeling language, BioNetGen, and the spatial extension, SRSim, have been developed. Here, we describe how to apply rule-based modeling to integrate experimental data from different sources into a single spatial simulation model and how to analyze the output of that model. The starting point for this approach can be a combination of molecular interaction data, reaction network data, proximities, binding and diffusion kinetics and molecular geometries at different levels of detail. We describe the technique and then use it to construct a model of the human mitotic inner and outer kinetochore, including the spindle assembly checkpoint signaling pathway. This allows us to demonstrate the utility of the procedure, show how a novel perspective for understanding such complex systems becomes accessible and elaborate on challenges that arise in the formulation, simulation and analysis of spatial rule-based models.

  2. Evolving MCDM applications using hybrid expert-based ISM and DEMATEL models: an example of sustainable ecotourism.

    Science.gov (United States)

    Chuang, Huan-Ming; Lin, Chien-Ku; Chen, Da-Ren; Chen, You-Shyang

    2013-01-01

    Ecological degradation is an escalating global threat. Increasingly, people are expressing awareness and priority for concerns about environmental problems surrounding them. Environmental protection issues are highlighted. An appropriate information technology tool, the growing popular social network system (virtual community, VC), facilitates public education and engagement with applications for existent problems effectively. Particularly, the exploration of related involvement behavior of VC member engagement is an interesting topic. Nevertheless, member engagement processes comprise interrelated sub-processes that reflect an interactive experience within VCs as well as the value co-creation model. To address the top-focused ecotourism VCs, this study presents an application of a hybrid expert-based ISM model and DEMATEL model based on multi-criteria decision making tools to investigate the complex multidimensional and dynamic nature of member engagement. Our research findings provide insightful managerial implications and suggest that the viral marketing of ecotourism protection is concerned with practitioners and academicians alike. PMID:24453902

  3. A hypergraph-based model for graph clustering: application to image indexing

    OpenAIRE

    Jouili, Salim; Tabbone, Salvatore

    2009-01-01

    Version finale disponible : www.springerlink.com International audience In this paper, we introduce a prototype-based clustering algorithm dealing with graphs. We propose a hypergraph-based model for graph data sets by allowing clusters overlapping. More precisely, in this representation one graph can be assigned to more than one cluster. Using the concept of the graph median and a given threshold, the proposed algorithm detects automatically the number of classes in the graph database....

  4. A new model for the grid size optimization of the finite element method --Based on its application to the water quality modeling of the topographically complicated river

    Institute of Scientific and Technical Information of China (English)

    ZENG Guangming; SU Xiaokang; HUANG Guohe; XIE Gengxin

    2003-01-01

    The finite element method is one of the typical methods that are used for numerical water quality modeling of the topographically complicated river. In this paper, based on the principle of probability theory the probability density of pollutants is introduced. A new model for the grid size optimization based on the finite element method is developed with the incorporation of the maximum information entropy theory when the length of the grid is given. Combined with the experiential evaluation approach of the flow discharge per unit river width, this model can be used to determine the grid size of the finite element method applied to water quality modeling of the topographically complicated river when the velocity field of the river is not given. The calculating results of the application of the model to an ideal river testified the correctness of the model. In a practical case-the application of the model to the Xingjian River (the Hengyang section of the Xiangjiang River), the optimized width of the grid of the finite element method was gained and the influence of parameters was studied, which demonstrated that the model reflected the real situation of the pollutants in the river, and that the model had many excellent characteristics such as stabilization, credibility and high applicability in practical applications.

  5. Application of nuclear models

    International Nuclear Information System (INIS)

    The development of extensive experimental nuclear data base over the past three decades has been accompanied by parallel advancement of nuclear theory and models used to describe and interpret the measurements. This theoretical capability is important because of many nuclear data requirements that are still difficult, impractical, or even impossible to meet with present experimental techniques. Examples of such data needs are neutron cross sections for unstable fission products, which are required for neutron absorption corrections in reactor calculations; cross sections for transactinide nuclei that control production of long-lived nuclear wastes; and the extensive dosimetry, activation, and neutronic data requirements to 40 MeV that must accompany development of the Fusion Materials Irradation Test (FMIT) facility. In recent years systematic improvements have been made in the nuclear models and codes used in data evaluation and, most importantly, in the methods used to derive physically based parameters for model calculations. The newly issued ENDF/B-V evaluated data library relies in many cases on nuclear reaction theory based on compound-nucleus Hauser-Feshbach, preequilibrium and direct reaction mechanisms as well as spherical and deformed optical-model theories. The development and applications of nuclear models for data evaluation are discussed with emphasis on the 1 to 40 MeV neutron energy range

  6. Physiologically Based Pharmacokinetic Modeling: Methodology, Applications, and Limitations with a Focus on Its Role in Pediatric Drug Development

    Directory of Open Access Journals (Sweden)

    Feras Khalil

    2011-01-01

    Full Text Available The concept of physiologically based pharmacokinetic (PBPK modeling was introduced years ago, but it has not been practiced significantly. However, interest in and implementation of this modeling technique have grown, as evidenced by the increased number of publications in this field. This paper demonstrates briefly the methodology, applications, and limitations of PBPK modeling with special attention given to discuss the use of PBPK models in pediatric drug development and some examples described in detail. Although PBPK models do have some limitations, the potential benefit from PBPK modeling technique is huge. PBPK models can be applied to investigate drug pharmacokinetics under different physiological and pathological conditions or in different age groups, to support decision-making during drug discovery, to provide, perhaps most important, data that can save time and resources, especially in early drug development phases and in pediatric clinical trials, and potentially to help clinical trials become more “confirmatory” rather than “exploratory”.

  7. Applications of sediment sudden deposition model based on the third-generation numerical model for shallow water wave

    Institute of Scientific and Technical Information of China (English)

    BAI Yuchuan; ZHANG Yinqi; ZHANG Bin

    2007-01-01

    The existing numerical models for nearshore waves are briefly introduced, and the third-generation numerical model for shallow water wave, which makes use of the most advanced productions of wave research and has been adapted well to be used in the environment of seacoast, lake and estuary area, is particularly discussed. The applied model realizes the significant wave height distribution at different wind directions. To integrate the model into the coastal area sediment, sudden deposition mechanism, the distribution of average silt content and the change of sediment sudden deposition thickness over time in the nearshore area are simulated. The academic productions can give some theoretical guidance to the applications of sediment sudden deposition mechanism for stormy waves in the coastal area. And the advancing directions of sediment sudden deposition model are prospected.

  8. The transient observation-based particle (TOP model and its potential application in radiation effects evaluation

    Directory of Open Access Journals (Sweden)

    Mazzino Laura

    2013-01-01

    Full Text Available The evaluation of the radiation hazards on components used in space environment is based on the knowledge of the radiation level encountered on orbit. The models that are widely used to assess the near-Earth environment for a given mission are empirical trapped radiation models derived from a compilation of spacecraft measurements. However, these models are static and hence are not suited for describing the short timescale variations of geomagnetic conditions. The transient observation-based particle (TOP-model tends to break with this classical approach by introducing dynamic features based on the observation and characterization of transient particle flux events in addition to classical mapping of steady-state flux levels. In order to get a preliminary version of an operational model (actually only available for electrons at low Earth orbit, LEO, (i the steady-state flux level, (ii the flux enhancements probability distribution functions, and (iii the flux decay-time constants (at given energy and positions in space were determined, and an original dynamic model skeleton with these input parameters has been developed. The methodology is fully described and first flux predictions from the model are presented. In order to evaluate the net effects of radiation on a component, it is important to have an efficient tool that calculates the transfer of the outer radiation environment through the spacecraft material, toward the location of the component under investigation. Using the TOP-model space radiation fluxes and the transmitted radiation environment characteristics derived through GEANT4 calculations, a case study for electron flux/dose variations in a small silicon volume is performed. Potential cases are assessed where the dynamic of the spacecraft radiation environment may have an impact on the observed radiation effects.

  9. Developing and Evaluating Creativity Gamification Rehabilitation System: The Application of PCA-ANFIS Based Emotions Model

    Science.gov (United States)

    Su, Chung-Ho; Cheng, Ching-Hsue

    2016-01-01

    This study aims to explore the factors in a patient's rehabilitation achievement after a total knee replacement (TKR) patient exercises, using a PCA-ANFIS emotion model-based game rehabilitation system, which combines virtual reality (VR) and motion capture technology. The researchers combine a principal component analysis (PCA) and an adaptive…

  10. ARX-NNPLS Model Based Optimization Strategy and Its Application in Polymer Grade Transition Process

    Institute of Scientific and Technical Information of China (English)

    费正顺; 胡斌; 叶鲁彬; 梁军

    2012-01-01

    Since it is often difficult to build differential algebraic equations (DAEs) for chemical processes, a new data-based modeling approach is proposed using ARX (AutoRegressive with eXogenous inputs) combined with neural network under partial least squares framework (ARX-NNPLS), in which less specific knowledge of the process is required but the input and output data. To represent the dynamic and nonlinear behavior of the process, the ARX combined with neural network is used in the partial least squares (PLS) inner model between input and output latent variables. In the proposed dynamic optimization strategy based on the ARX-NNPLS model, neither parameterization nor iterative solving process for DAEs is needed as the ARX-NNPLS model gives a proper representation for the dynamic behavior of the process, and the computing time is greatly reduced compared to conventional control vector parameterization method. To demonstrate the ARX-NNPLS model based optimization strategy, the polyethylene grade transition in gas phase fluidized-bed reactor is taken into account. The optimization results show that the final optimal trajectory of quality index determined by the new approach moves faster to the target values and the computing time is much less.

  11. Function Based Nonlinear Least Squares and Application to Jelinski--Moranda Software Reliability Model

    CERN Document Server

    Liu, Jingwei

    2011-01-01

    A function based nonlinear least squares estimation (FNLSE) method is proposed and investigated in parameter estimation of Jelinski-Moranda software reliability model. FNLSE extends the potential fitting functions of traditional least squares estimation (LSE), and takes the logarithm transformed nonlinear least squares estimation (LogLSE) as a special case. A novel power transformation function based nonlinear least squares estimation (powLSE) is proposed and applied to the parameter estimation of Jelinski-Moranda model. Solved with Newton-Raphson method, Both LogLSE and powLSE of Jelinski-Moranda models are applied to the mean time between failures (MTBF) predications on six standard software failure time data sets. The experimental results demonstrate the effectiveness of powLSE with optimal power index compared to the classical least--squares estimation (LSE), maximum likelihood estimation (MLE) and LogLSE in terms of recursively relative error (RE) index and Braun statistic index.

  12. Improvements in mode-based waveform modeling and application to Eurasian velocity structure

    Science.gov (United States)

    Panning, M. P.; Marone, F.; Kim, A.; Capdeville, Y.; Cupillard, P.; Gung, Y.; Romanowicz, B.

    2006-12-01

    We introduce several recent improvements to mode-based 3D and asymptotic waveform modeling and examine how to integrate them with numerical approaches for an improved model of upper-mantle structure under eastern Eurasia. The first step in our approach is to create a large-scale starting model including shear anisotropy using Nonlinear Asymptotic Coupling Theory (NACT; Li and Romanowicz, 1995), which models the 2D sensitivity of the waveform to the great-circle path between source and receiver. We have recently improved this approach by implementing new crustal corrections which include a non-linear correction for the difference between the average structure of several large regions from the global model with further linear corrections to account for the local structure along the path between source and receiver (Marone and Romanowicz, 2006; Panning and Romanowicz, 2006). This model is further refined using a 3D implementation of Born scattering (Capdeville, 2005). We have made several recent improvements to this method, in particular introducing the ability to represent perturbations to discontinuities. While the approach treats all sensitivity as linear perturbations to the waveform, we have also experimented with a non-linear modification analogous to that used in the development of NACT. This allows us to treat large accumulated phase delays determined from a path-average approximation non-linearly, while still using the full 3D sensitivity of the Born approximation. Further refinement of shallow regions of the model is obtained using broadband forward finite-difference waveform modeling. We are also integrating a regional Spectral Element Method code into our tomographic modeling, allowing us to move beyond many assumptions inherent in the analytic mode-based approaches, while still taking advantage of their computational efficiency. Illustrations of the effects of these increasingly sophisticated steps will be presented.

  13. A Patch-Based Structural Masking Model with an Application to Compression

    Directory of Open Access Journals (Sweden)

    Damon M. Chandler

    2009-01-01

    Full Text Available The ability of an image region to hide or mask a given target signal continues to play a key role in the design of numerous image processing and vision systems. However, current state-of-the-art models of visual masking have been optimized for artificial targets placed upon unnatural backgrounds. In this paper, we (1 measure the ability of natural-image patches in masking distortion; (2 analyze the performance of a widely accepted standard masking model in predicting these data; and (3 report optimal model parameters for different patch types (textures, structures, and edges. Our results reveal that the standard model of masking does not generalize across image type; rather, a proper model should be coupled with a classification scheme which can adapt the model parameters based on the type of content contained in local image patches. The utility of this adaptive approach is demonstrated via a spatially adaptive compression algorithm which employs patch-based classification. Despite the addition of extra side information and the high degree of spatial adaptivity, this approach yields an efficient wavelet compression strategy that can be combined with very accurate rate-control procedures.

  14. A Patch-Based Structural Masking Model with an Application to Compression

    Directory of Open Access Journals (Sweden)

    Gaubatz MatthewD

    2009-01-01

    Full Text Available Abstract The ability of an image region to hide or mask a given target signal continues to play a key role in the design of numerous image processing and vision systems. However, current state-of-the-art models of visual masking have been optimized for artificial targets placed upon unnatural backgrounds. In this paper, we (1 measure the ability of natural-image patches in masking distortion; (2 analyze the performance of a widely accepted standard masking model in predicting these data; and (3 report optimal model parameters for different patch types (textures, structures, and edges. Our results reveal that the standard model of masking does not generalize across image type; rather, a proper model should be coupled with a classification scheme which can adapt the model parameters based on the type of content contained in local image patches. The utility of this adaptive approach is demonstrated via a spatially adaptive compression algorithm which employs patch-based classification. Despite the addition of extra side information and the high degree of spatial adaptivity, this approach yields an efficient wavelet compression strategy that can be combined with very accurate rate-control procedures.

  15. REST based mobile applications

    Science.gov (United States)

    Rambow, Mark; Preuss, Thomas; Berdux, Jörg; Conrad, Marc

    2008-02-01

    Simplicity is the major advantage of REST based webservices. Whereas SOAP is widespread in complex, security sensitive business-to-business aplications, REST is widely used for mashups and end-user centric applicatons. In that context we give an overview of REST and compare it to SOAP. Furthermore we apply the GeoDrawing application as an example for REST based mobile applications and emphasize on pros and cons for the use of REST in mobile application scenarios.

  16. The Application of Community-Based Economic Model in Entrepreneurial Activities

    Directory of Open Access Journals (Sweden)

    Sukristin Sukristin

    2014-12-01

    Full Text Available One of the efforts to increase people's income is by applying economic models that are suitable for the community. This model places emphasis on lower class community economic empowerment so that eventually they can be economically self-sufficient. To apply an economy model suitable for that particular source, around the area of Sungai Melayu, particularly Piansak village is one of the efforts suggested. The method applied for this research is Action Research. For the first year, the main task is to identify the local area. Only after a feasibility study being done on the data found, then a system of Society Economy Model will be implemented accordingly. The further steps will follow based on the result yielded of first-year-research. The potential businesses suitable to the area, possibly will be businesses in raising livestock or poultry, such as

  17. Microstructurally based thermomechanical fatigue lifetime model of solder joints for electronic applications

    Energy Technology Data Exchange (ETDEWEB)

    Frear, D.R.; Rashid, M.M.; Burchett, S.N.

    1993-07-01

    We present a new methodology for predicting the fatigue life of solder joints for electronics applications. This approach involves integration of experimental and computational techniques. The first stage involves correlating the manufacturing and processing parameters with the starting microstructure of the solder joint. The second stage involves a series of experiments that characterize the evolution of the microstructure during thermal cycling. The third stage consists of a computer modeling and simulation effort that utilizes the starting microstructure and experimental data to produce a reliability prediction of the solder joint. This approach is an improvement over current methodologies because it incorporates the microstructure and properties of the solder directly into the model and allows these properties to evolve as the microstructure changes during fatigue.

  18. Tomogram-based comparison of geostatistical models: Application to the Macrodispersion Experiment (MADE) site

    Science.gov (United States)

    Linde, Niklas; Lochbühler, Tobias; Dogan, Mine; Van Dam, Remke L.

    2015-12-01

    We propose a new framework to compare alternative geostatistical descriptions of a given site. Multiple realizations of each of the considered geostatistical models and their corresponding tomograms (based on inversion of noise-contaminated simulated data) are used as a multivariate training image. The training image is scanned with a direct sampling algorithm to obtain conditional realizations of hydraulic conductivity that are not only in agreement with the geostatistical model, but also honor the spatially varying resolution of the site-specific tomogram. Model comparison is based on the quality of the simulated geophysical data from the ensemble of conditional realizations. The tomogram in this study is obtained by inversion of cross-hole ground-penetrating radar (GPR) first-arrival travel time data acquired at the MAcro-Dispersion Experiment (MADE) site in Mississippi (USA). Various heterogeneity descriptions ranging from multi-Gaussian fields to fields with complex multiple-point statistics inferred from outcrops are considered. Under the assumption that the relationship between porosity and hydraulic conductivity inferred from local measurements is valid, we find that conditioned multi-Gaussian realizations and derivatives thereof can explain the crosshole geophysical data. A training image based on an aquifer analog from Germany was found to be in better agreement with the geophysical data than the one based on the local outcrop, which appears to under-represent high hydraulic conductivity zones. These findings are only based on the information content in a single resolution-limited tomogram and extending the analysis to tracer or higher resolution surface GPR data might lead to different conclusions (e.g., that discrete facies boundaries are necessary). Our framework makes it possible to identify inadequate geostatistical models and petrophysical relationships, effectively narrowing the space of possible heterogeneity representations.

  19. Risk Evaluation Approach and Application Research on Fuzzy-FMECA Method Based on Cloud Model

    Directory of Open Access Journals (Sweden)

    Zhengjie Xu

    2013-09-01

    Full Text Available In order to safeguard the safety of passengers and reducemaintenance costs, it is necessary to analyze and evaluate the security risk ofthe Railway Signal System. However, the conventional Fuzzy Analytical HierarchyProcess (FAHP can not describe the fuzziness and randomness of the judgment,accurately, and once the fuzzy sets are described using subjection degreefunction, the concept of fuzziness will be no longer fuzzy. Thus Fuzzy-FMECAmethod based on cloud model is put forward. Failure Modes Effects andCriticality Analysis (FMECA method is used to identify the risk and FAHP basedon cloud model is used for determining the subjection degree function in fuzzymethod, finally the group decision can be gained with the syntheticallyaggregated cloud model, the method’s feasibility and effectiveness are shown inthe practical examples. Finally Fuzzy-FMECA based on cloud model and theconventional FAHP are used to assess the risk respectively, evaluation resultsshow that the cloud model which is introduced into the risk assessment ofRailway Signal System can realize the transition between precise value andquality value by combining the fuzziness and randomness and provide moreabundant information than subjection degree function of the conventional FAHP.

  20. Modeling watershed-scale solute transport using an integrated, process-based hydrologic model with applications to bacterial fate and transport

    Science.gov (United States)

    Niu, Jie; Phanikumar, Mantha S.

    2015-10-01

    Distributed hydrologic models that simulate fate and transport processes at sub-daily timescales are useful tools for estimating pollutant loads exported from watersheds to lakes and oceans downstream. There has been considerable interest in the application of integrated process-based hydrologic models in recent years. While the models have been applied to address questions of water quantity and to better understand linkages between hydrology and land surface processes, routine applications of these models to address water quality issues are currently limited. In this paper, we first describe a general process-based watershed-scale solute transport modeling framework, based on an operator splitting strategy and a Lagrangian particle transport method combined with dispersion and reactions. The transport and the hydrologic modules are tightly coupled and the interactions among different hydrologic components are explicitly modeled. We test transport modules using data from plot-scale experiments and available analytical solutions for different hydrologic domains. The numerical solutions are also compared with an analytical solution for groundwater transit times with interactions between surface and subsurface flows. Finally, we demonstrate the application of the model to simulate bacterial fate and transport in the Red Cedar River watershed in Michigan and test hypotheses about sources and transport pathways. The watershed bacterial fate and transport model is expected to be useful for making near real-time predictions at marine and freshwater beaches.

  1. Is equine colic seasonal? Novel application of a model based approach

    Directory of Open Access Journals (Sweden)

    Proudman Christopher J

    2006-08-01

    Full Text Available Abstract Background Colic is an important cause of mortality and morbidity in domesticated horses yet many questions about this condition remain to be answered. One such question is: does season have an effect on the occurrence of colic? Time-series analysis provides a rigorous statistical approach to this question but until now, to our knowledge, it has not been used in this context. Traditional time-series modelling approaches have limited applicability in the case of relatively rare diseases, such as specific types of equine colic. In this paper we present a modelling approach that respects the discrete nature of the count data and, using a regression model with a correlated latent variable and one with a linear trend, we explored the seasonality of specific types of colic occurring at a UK referral hospital between January 1995–December 2004. Results Six- and twelve-month cyclical patterns were identified for all colics, all medical colics, epiploic foramen entrapment (EFE, equine grass sickness (EGS, surgically treated and large colon displacement/torsion colic groups. A twelve-month cyclical pattern only was seen in the large colon impaction colic group. There was no evidence of any cyclical pattern in the pedunculated lipoma group. These results were consistent irrespective of whether we were using a model including latent correlation or trend. Problems were encountered in attempting to include both trend and latent serial dependence in models simultaneously; this is likely to be a consequence of a lack of power to separate these two effects in the presence of small counts, yet in reality the underlying physical effect is likely to be a combination of both. Conclusion The use of a regression model with either an autocorrelated latent variable or a linear trend has allowed us to establish formally a seasonal component to certain types of colic presented to a UK referral hospital over a 10 year period. These patterns appeared to coincide

  2. Alternatives to accuracy and bias metrics based on percentage errors for radiation belt modeling applications

    Energy Technology Data Exchange (ETDEWEB)

    Morley, Steven Karl [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-07-01

    This report reviews existing literature describing forecast accuracy metrics, concentrating on those based on relative errors and percentage errors. We then review how the most common of these metrics, the mean absolute percentage error (MAPE), has been applied in recent radiation belt modeling literature. Finally, we describe metrics based on the ratios of predicted to observed values (the accuracy ratio) that address the drawbacks inherent in using MAPE. Specifically, we define and recommend the median log accuracy ratio as a measure of bias and the median symmetric accuracy as a measure of accuracy.

  3. Research and application of mineral resources assessment by weights of evidence model based on SIG

    Institute of Scientific and Technical Information of China (English)

    Yuanyuan Chuai; Keyan Xiao; Yihua Xuan; Shaobin Zhan

    2006-01-01

    Geological data are usually of the characteristics of multi-source, large amount and multi-scale. The construction of Spatial Information Grid overcomes the shortages of personal computers when dealing with geological data. The authors introduce the definition, architecture and flow of mineral resources assessment by weights of evidence model based on Spatial Information Grid (SIG). Meanwhile, a case study on the prediction of copper mineral occurrence in the Middle-Lower Yangtze metallogenic belt is given. The results show that mineral resources assessement based on SIG is an effective new method which provides a way of sharing and integrating distributed geospatial information and improves the efficiency greatly.

  4. Alternatives to accuracy and bias metrics based on percentage errors for radiation belt modeling applications

    Energy Technology Data Exchange (ETDEWEB)

    Morley, Steven Karl [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2016-07-01

    This report reviews existing literature describing forecast accuracy metrics, concentrating on those based on relative errors and percentage errors. We then review how the most common of these metrics, the mean absolute percentage error (MAPE), has been applied in recent radiation belt modeling literature. Finally, we describe metrics based on the ratios of predicted to observed values (the accuracy ratio) that address the drawbacks inherent in using MAPE. Specifically we define and recommend the median log accuracy ratio as a measure of bias and the median symmetric accuracy as a measure of accuracy.

  5. Radioactive Threat Detection with Scattering Physics: A Model-Based Application

    Energy Technology Data Exchange (ETDEWEB)

    Candy, J V; Chambers, D H; Breitfeller, E F; Guidry, B L; Verbeke, J M; Axelrod, M A; Sale, K E; Meyer, A M

    2010-01-21

    The detection of radioactive contraband is a critical problem in maintaining national security for any country. Emissions from threat materials challenge both detection and measurement technologies especially when concealed by various types of shielding complicating the transport physics significantly. The development of a model-based sequential Bayesian processor that captures both the underlying transport physics including scattering offers a physics-based approach to attack this challenging problem. It is shown that this processor can be used to develop an effective detection technique.

  6. A PV/T and Heat Pump based trigeneration system model for residential applications

    OpenAIRE

    Joyce, António; Coelho, Luis; Martins, João F.; Tavares, Nelson; R Pereira; Magalhães, Pedro

    2011-01-01

    A solar trigeneration system, based on photovoltaic-thermal (PV/T) collectors, photovoltaic (PV) modules and a heat pump unit for heating and cooling, is modelled to forecast the thermal and electric yields of the system. The aim of the trigeneration system is to provide enough electricity, domestic hot water (DHW), heating and cooling power to meet the typical demand of an urban single family dwelling with limited roof area and allow the household to achieve a positive net energy status. The...

  7. Model based fault diagnosis for hybrid systems : application on chemical processes

    OpenAIRE

    Olivier Maget, Nelly; Hétreux, Gilles; Le Lann, Jean-Marc

    2009-01-01

    The complexity and the size of the industrial chemical processes induce the monitoring of a growing number of process variables. Their knowledge is generally based on the measurements of system variables and on the physico-chemical models of the process. Nevertheless, this information is imprecise because of process and measurement noise. So the research ways aim at developing new and more powerful techniques for the detection of process fault. In this work, we present a method for the fault ...

  8. Model-based Measures of Output Gap: Application to the Thai Economy

    OpenAIRE

    Vimut Vanitcharearnthum

    2012-01-01

    In this paper we compare two model-based measures of the output gap. The first measure, as proposed by Gali (2011), defines output gap as the difference between actual output and the output level that would be if the economy operates under a perfectly competitive market without price or wage stickiness. We used annual data of relevant variables for Thailand and computed the output gap under this approach. The calculated output gap for Thailand shows that the Thai economy performs consistently...

  9. An engineering process for security patterns application in component based models

    OpenAIRE

    Bouaziz, Rahma; Kallel, Slim; Coulette, Bernard

    2013-01-01

    Security engineering with patterns is currently a very active area of research. Security patterns - an adaptation of Design Patterns to security - capture experts' experience in order to solve recurrent security problems in a structured and reusable way. In this paper, our objective is to describe an engineering process, called SCRIP (SeCurity patteRn Integration Process), which provides guidelines for integrating security patterns into component-based models. SCRIP defines activities and pro...

  10. Model based fault diagnosis in a centrifugal pump application using structural analysis

    DEFF Research Database (Denmark)

    Kallesøe, C. S.; Izadi-Zamanabadi, Roozbeh; Rasmussen, Henrik;

    2004-01-01

    A model based approach for fault detection and isolation in a centrifugal pump is proposed in this paper. The fault detection algorithm is derived using a combination of structural analysis, Analytical Redundant Relations (ARR) and observer designs. Structural considerations on the system are used...... to an industrial benchmark. The benchmark tests have shown that the algorithm is capable of detection and isolation of five different faults in the mechanical and hydraulic parts of the pump....

  11. Model Based Fault Diagnosis in a Centrifugal Pump Application using Structural Analysis

    DEFF Research Database (Denmark)

    Kallesøe, C. S.; Izadi-Zamanabadi, Roozbeh; Rasmussen, Henrik;

    2004-01-01

    A model based approach for fault detection and isolation in a centrifugal pump is proposed in this paper. The fault detection algorithm is derived using a combination of structural analysis, Analytical Redundant Relations (ARR) and observer designs. Structural considerations on the system are used...... to an industrial benchmark. The benchmark tests have shown that the algorithm is capable of detection and isolation of five different faults in the mechanical and hydraulic parts of the pump....

  12. [The application of cybernetic modeling methods for the forensic medical personality identification based on the voice and sounding speech characteristics].

    Science.gov (United States)

    Kaganov, A Sh; Kir'yanov, P A

    2015-01-01

    The objective of the present publication was to discuss the possibility of application of cybernetic modeling methods to overcome the apparent discrepancy between two kinds of the speech records, viz. initial ones (e.g. obtained in the course of special investigation activities) and the voice prints obtained from the persons subjected to the criminalistic examination. The paper is based on the literature sources and the materials of original criminalistics expertises performed by the authors. PMID:26245103

  13. LINKING SATELLITE REMOTE SENSING BASED ENVIRONMENTAL PREDICTORS TO DISEASE: AN APPLICATION TO THE SPATIOTEMPORAL MODELLING OF SCHISTOSOMIASIS IN GHANA

    OpenAIRE

    Wrable, M.; Liss, A.; Kulinkina, A.; Koch, M.; Biritwum, N. K.; Ofosu, A.; Kosinski, K. C.; Gute, D M; Naumova, E. N.

    2016-01-01

    90% of the worldwide schistosomiasis burden falls on sub-Saharan Africa. Control efforts are often based on infrequent, small-scale health surveys, which are expensive and logistically difficult to conduct. Use of satellite imagery to predictively model infectious disease transmission has great potential for public health applications. Transmission of schistosomiasis requires specific environmental conditions to sustain freshwater snails, however has unknown seasonality, and is difficult to s...

  14. Engine Modelling for Control Applications

    DEFF Research Database (Denmark)

    Hendricks, Elbert

    1997-01-01

    In earlier work published by the author and co-authors, a dynamic engine model called a Mean Value Engine Model (MVEM) was developed. This model is physically based and is intended mainly for control applications. In its newer form, it is easy to fit to many different engines and requires little ...

  15. A microcantilever-based alcohol vapor sensor-application and response model

    DEFF Research Database (Denmark)

    Jensenius, Henriette; Thaysen, Jacob; Rasmussen, Anette Alsted; Veje, Lars Helt; Hansen, Ole; Boisen, Anja

    2000-01-01

    A recently developed microcantilever probe with integrated piezoresistive readout has been applied as a gas sensor. Resistors, sensitive to stress changes, are integrated on the flexible cantilevers. This makes it possible to monitor the cantilever deflection electrically and with an integrated...... reference cantilever background noise is subtracted directly in the measurement. A polymer coated cantilever has been exposed to vapors of various alcohols and the resulting cantilever response has been interpreted using a simple evaporation model. The model indicates that the cantilever response is a...... direct measure of the molecular concentration of alcohol vapor. On the basis of the model the detection limit of this cantilever-based sensor is determined to be below 10 ppm for alcohol vapor measurements. Furthermore, the time response of the cantilever can be used to distinguish between different...

  16. Modeling of Soft sensor based on Artificial Neural Network for Galactic Cosmic Rays Application

    International Nuclear Information System (INIS)

    For successful designing of space radiation Galactic Cosmic Rays (GCRs) model, we develop a soft sensor based on the Artificial Neural Network (ANN) model. At the first step, the soft sensor based ANN was constructed as an alternative to model space radiation environment. The structure of ANN in this model is using Multilayer Perceptron (MLP) and Levenberg Marquardt algorithms with 3 inputs and 2 outputs. In the input variable, we use 12 years data (Corr, Uncorr and Press) of GCR particles obtained from Neutron Monitor of Bartol University (Fort Smith area) and the target output is (Corr and Press) from the same source but for Inuvik area in the Polar Regions. In the validation step, we obtained the Root Mean Square Error (RMSE) value of Corr 3.8670e-004 and Press 1.3414e-004 and Variance Accounted For (VAF) of Corr 99.9839 % and Press 99.9831% during the training section. After all the results obtained, then we applied into a Matlab GUI simulation (soft sensor simulation). This simulation will display the estimation of output value from input (Corr and Press). Testing results showed an error of 0.133% and 0.014% for Corr and Press, respectively

  17. Monitoring of rainfall by ground-based passive microwave systems: models, measurements and applications

    Directory of Open Access Journals (Sweden)

    F. S. Marzano

    2005-01-01

    Full Text Available A large set of ground-based multi-frequency microwave radiometric simulations and measurements during different precipitation regimes are analysed. Simulations are performed for a set of frequencies from 22 to 60 GHz, representing the channels currently available on an operational ground-based radiometric system. Results are illustrated in terms of comparisons between measurements and model data in order to show that the observed radiometric signatures can be attributed to rainfall scattering and absorption. An inversion algorithm has been developed, basing on the simulated data, to retrieve rain rate from passive radiometric observations. As a validation of the approach, we have analyzed radiometric measurements during rain events occurred in Boulder, Colorado, and at the Atmospheric Radiation Measurement (ARM Program's Southern Great Plains (SGP site in Lamont, Oklahoma, USA, comparing rain rate estimates with available simultaneous rain gauge data.

  18. Application of a circulation model in waters, based in the difference method, for bays

    International Nuclear Information System (INIS)

    The Knowledge of circulation of water in bays, in addition to the possibility of simulation future conditions, can be of great interest in solving problems related to the cooling water for Nuclear Power Plants, study of sediments and water polution, in addition to the study of civil engineering works planned in bays. A Numerical Circulation Model of water in bays, is applied to the conditions of Sepetiba Bay at Rio de Janeiro coast. This System of Partial Differential Equations that constitute the Model, were solved by the Finite Difference Method, using a uniform cartesian grid for uniform time steps generating a bi-dimensional flow measurement of depth. The results obtained by comparing the values of the Model and measurements taken a bay were satisfactory, assuring its credibility and efficiency. A programming code was developed for the application providing outputing at any preditermined time steps, with discrimination of 30 seconds, the average levels, flows, velocities and depths of water of each grid spacing along the length of the bay in addition to a graphic of the flow. (Author)

  19. Pseudo-emulsion based hollow fiber strip dispersion technique modeling and application for Pu (IV) recovery from oxalate supernatant waste

    International Nuclear Information System (INIS)

    The extraction - recovery of Pu(IV) from nitric acidic medium has been examined by PEHFSD (Pseudo-emulsion based hollow fiber strip dispersion) in a microporous hydrophobic polypropylene hollow fiber membrane contactor. The technology of supported liquid membranes with strip dispersion has advantages over the conventional supported liquid membranes solving the inherent stability problems of the membrane. Mathematical models for extraction in hollow-fiber contactor were also developed. The mass transfer coefficients were calculated from the experimental results and models are presented. The applicability of PEHFSD technique was demonstrated for recovery of Pu(IV) from oxalate supernatant waste generated during plutonium precipitation by oxalic acid. (author)

  20. Application of model-based spectral analysis to wind-profiler radar observations

    Directory of Open Access Journals (Sweden)

    E. Boyer

    Full Text Available A classical way to reduce a radar’s data is to compute the spectrum using FFT and then to identify the different peak contributions. But in case an overlapping between the different echoes (atmospheric echo, clutter, hydrometeor echo. . . exists, Fourier-like techniques provide poor frequency resolution and then sophisticated peak-identification may not be able to detect the different echoes. In order to improve the number of reduced data and their quality relative to Fourier spectrum analysis, three different methods are presented in this paper and applied to actual data. Their approach consists of predicting the main frequency-components, which avoids the development of very sophisticated peak-identification algorithms. The first method is based on cepstrum properties generally used to determine the shift between two close identical echoes. We will see in this paper that this method cannot provide a better estimate than Fourier-like techniques in an operational use. The second method consists of an autoregressive estimation of the spectrum. Since the tests were promising, this method was applied to reduce the radar data obtained during two thunder-storms. The autoregressive method, which is very simple to implement, improved the Doppler-frequency data reduction relative to the FFT spectrum analysis. The third method exploits a MUSIC algorithm, one of the numerous subspace-based methods, which is well adapted to estimate spectra composed of pure lines. A statistical study of performances of this method is presented, and points out the very good resolution of this estimator in comparison with Fourier-like techniques. Application to actual data confirms the good qualities of this estimator for reducing radar’s data.

    Key words. Meteorology and atmospheric dynamics (tropical meteorology- Radio science (signal processing- General (techniques applicable in three or more fields

  1. New Combined Weighting Model Based on Maximizing the Difference in Evaluation Results and Its Application

    Directory of Open Access Journals (Sweden)

    Bin Meng

    2015-01-01

    Full Text Available This paper presents an approach for weighting indices in the comprehensive evaluation. In accordance with the principle that the entire difference of various evaluation objects is to be maximally differentiated, an adjusted weighting coefficient is introduced. Based on the idea of maximizing the difference between the adjusted evaluation scores of each evaluation object and their mean, an objective programming model is established with more obvious differentiation between evaluation scores and the combined weight coefficient determined, thereby avoiding contradictory and less distinguishable evaluation results of single weighting methods. The proposed model is demonstrated using 2,044 observations. The empirical results show that the combined weighting method has the least misjudgment probability, as well as the least error probability, when compared with four single weighting methods, namely, G1, G2, variation coefficient, and deviation methods.

  2. A Common Reasoning Model and Its Application in Knowledge—Based System

    Institute of Scientific and Technical Information of China (English)

    郑方青

    1991-01-01

    To use reasoning knowledge accurately and efficiently,many reasoning methods have been proposed.However,the differences in form among the methods may obstruct the systematical analysis and harmonious integration of them.In this paper,a common reasoning model JUM(Judgement Model)is introduced.According to JUM,a common knowledge representation form is abstracted from different reasoning methods and its limitation is reduced.We also propose an algorithm for transforming one type of JUMs into another.In some cases,the algorithm can be used to resolve the key problem of integrating different types of JUM in one system.It is possible that a new architecture of knowledge-based system can be realized under JUM.

  3. A resonance-based solar element: a numerical model and micro/nano technology application

    Science.gov (United States)

    Fiala, P.; Drexler, P.; Nespor, D.

    2013-05-01

    The authors propose an analysis of a model solar element based on the principle of a resonance system facilitating the transformation of the external form of incident energy into electrical energy. A similar principle provides the basis for harvesters designed to operate at lower frequencies, Jirků T., Fiala P. and Kluge M.,2010, Wen J.L., Wen Z., Wong P.K., 2000. In these harvesters, the efficiency of the energy form transformation can be controlled from the frequency spectrum of an external source (the Sun).

  4. A Parallel Decision Model Based on Support Vector Machines and Its Application to Fault Diagnosis

    Institute of Scientific and Technical Information of China (English)

    Yan Weiwu(阎威武); Shao Huihe

    2004-01-01

    Many industrial process systems are becoming more and more complex and are characterized by distributed features. To ensure such a system to operate under working order, distributed parameter values are often inspected from subsystems or different points in order to judge working conditions of the system and make global decisions. In this paper, a parallel decision model based on Support Vector Machine (PDMSVM) is introduced and applied to the distributed fault diagnosis in industrial process. PDMSVM is convenient for information fusion of distributed system and it performs well in fault diagnosis with distributed features. PDMSVM makes decision based on synthetic information of subsystems and takes the advantage of Support Vector Machine. Therefore decisions made by PDMSVM are highly reliable and accurate.

  5. Model-Based Reinforcement of Kinect Depth Data for Human Motion Capture Applications

    Directory of Open Access Journals (Sweden)

    Andreas Skiadopoulos

    2013-07-01

    Full Text Available Motion capture systems have recently experienced a strong evolution. New cheap depth sensors and open source frameworks, such as OpenNI, allow for perceiving human motion on-line without using invasive systems. However, these proposals do not evaluate the validity of the obtained poses. This paper addresses this issue using a model-based pose generator to complement the OpenNI human tracker. The proposed system enforces kinematics constraints, eliminates odd poses and filters sensor noise, while learning the real dimensions of the performer’s body. The system is composed by a PrimeSense sensor, an OpenNI tracker and a kinematics-based filter and has been extensively tested. Experiments show that the proposed system improves pure OpenNI results at a very low computational cost.

  6. A Critical Review on Wind Turbine Power Curve Modelling Techniques and Their Applications in Wind Based Energy Systems

    Directory of Open Access Journals (Sweden)

    Vaishali Sohoni

    2016-01-01

    Full Text Available Power curve of a wind turbine depicts the relationship between output power and hub height wind speed and is an important characteristic of the turbine. Power curve aids in energy assessment, warranty formulations, and performance monitoring of the turbines. With the growth of wind industry, turbines are being installed in diverse climatic conditions, onshore and offshore, and in complex terrains causing significant departure of these curves from the warranted values. Accurate models of power curves can play an important role in improving the performance of wind energy based systems. This paper presents a detailed review of different approaches for modelling of the wind turbine power curve. The methodology of modelling depends upon the purpose of modelling, availability of data, and the desired accuracy. The objectives of modelling, various issues involved therein, and the standard procedure for power performance measurement with its limitations have therefore been discussed here. Modelling methods described here use data from manufacturers’ specifications and actual data from the wind farms. Classification of modelling methods, various modelling techniques available in the literature, model evaluation criteria, and application of soft computing methods for modelling are then reviewed in detail. The drawbacks of the existing methods and future scope of research are also identified.

  7. Forest dynamics in Oregon landscapes: Evaluation and application of an individual-based model

    Science.gov (United States)

    Busing, R.T.; Solomon, A.M.; McKane, R.B.; Burdick, C.A.

    2007-01-01

    The FORCLIM model of forest dynamics was tested against field survey data for its ability to simulate basal area and composition of old forests across broad climatic gradients in western Oregon, USA. The model was also tested for its ability to capture successional trends in ecoregions of the west Cascade Range. It was then applied to simulate present and future (1990-2050) forest landscape dynamics of a watershed in the west Cascades. Various regimes of climate change and harvesting in the watershed were considered in the landscape application. The model was able to capture much of the variation in forest basal area and composition in western Oregon even though temperature and precipitation were the only inputs that were varied among simulated sites. The measured decline in total basal area from tall coastal forests eastward to interior steppe was matched by simulations. Changes in simulated forest dominants also approximated those in the actual data. Simulated abundances of a few minor species did not match actual abundances, however. Subsequent projections of climate change and harvest effects in a west Cascades landscape indicated no change in forest dominance as of 2050. Yet, climate-driven shifts in the distributions of some species were projected. The simulation of both stand-replacing and partial-stand disturbances across western Oregon improved agreement between simulated and actual data. Simulations with fire as an agent of partial disturbance suggested that frequent fires of low severity can alter forest composition and structure as much or more than severe fires at historic frequencies. ?? 2007 by the Ecological Society of America.

  8. MAIN REGULARITIES OF FAULTING IN LITHOSPHERE AND THEIR APPLICATION (BASED ON PHYSICAL MODELLING RESULTS

    Directory of Open Access Journals (Sweden)

    S. A. Bornyakov

    2015-09-01

    Full Text Available Results of long-term experimental studies and modelling of faulting are briefly reviewed, and research methods and the-state-of-art issues are described. The article presents the main results of faulting modelling with the use of non-transparent elasto-viscous plastic and optically active models. An area of active dynamic influence of fault (AADIF is the term introduced to characterise a fault as a 3D geological body. It is shown that AADIF's width (М is determined by thickness of the layer wherein a fault occurs (Н, its viscosity (η and strain rate (V. Multiple correlation equations are proposed to show relationships between AADIF's width (М, H, η and V for faults of various morphological and genetic types. The irregularity of AADIF in time and space is characterised in view of staged formation of the internal fault structure of such areas and geometric and dynamic parameters of AADIF which are changeable along the fault strike. The authors pioneered in application of the open system conception to find explanations of regularities of structure formation in AADIFs. It is shown that faulting is a synergistic process of continuous changes of structural levels of strain, which differ in manifestation of specific self-similar fractures of various scales. Such levels are changeable due to self-organization processes of fracture systems. Fracture dissipative structures (FDS is the term introduced to describe systems of fractures that are subject to self-organization. It is proposed to consider informational entropy and fractal dimensions in order to reveal FDS in AADIF. Studied are relationships between structure formation in AADIF and accompanying processes, such as acoustic emission and terrain development above zones wherein faulting takes place. Optically active elastic models were designed to simulate the stress-and-strain state of AADIF of main standard types of fault jointing zones and their analogues in nature, and modelling results are

  9. Reconstruction of adsorption potential in Polanyi-based models and application to various adsorbents.

    Science.gov (United States)

    Pan, Bingjun; Zhang, Huichun

    2014-06-17

    The equilibrium Polanyi adsorption potential was reconstructed as ε = -RT ln(Ca(or H)/δ) to correlate the characteristic energy (E) of Polanyi-based models (qe = f[ε/E]) with the properties or structures of absorbates, where qe is the equilibriumn adsorption capacity, Ca(or H) is the converted concentration from the equilibrium aqueous concentration at the same activity and corresponds to the adsorption from the gas or n-hexadecane (HD) phase by the water-wet adsorbent, and "δ" is an arbitrary divisor to converge the model fitting. Subsequently, the modified Dubinin-Astakhov model based on the reconstructed ε was applied to aqueous adsorption on activated carbon, black carbon, multiwalled carbon nanotubes, and polymeric resin. The fitting results yielded intrinsic characteristic energies Ea, derived from aqueous-to-gas phase conversion, or EH, derived from aqueous-to-HD phase conversion, which reflect the contributions of the overall or specific adsorbate-adsorbent interactions to the adsorption. Effects of the adsorbate and adsorbent properties on Ea or EH then emerge that are unrevealed by the original characteristic energy (Eo), i.e., adsorbates with tendency to form stronger interactions with an adsorbent have larger Ea and EH. Additionally, comparison of Ea and EH allows quantitative analysis of the contributions of nonspecific interactions, that is, a significant relationship was established between the nonspecific interactions and Abraham's descriptors for the adsorption of all 32 solutes on the four different adsorbents: (Ea - EH) = 24.7 × V + 9.7 × S - 19.3 (R(2) = 0.97), where V is McGowan's characteristic volume for adsorbates, and S reflects the adsorbate's polarity/polarizability. PMID:24815932

  10. APPLICATION OF A MODIFIED QUICK SCHEME TO DEPTHAVERAGED k-( TURBULENCE MODEL BASED ON UNSTRUCTURED GRIDS

    Institute of Scientific and Technical Information of China (English)

    HUA Zu-lin; XING Ling-hang; GU Li

    2008-01-01

    The modified QUICK scheme on unstructured grid was used to improve the advection flux approximation, and the depth-averaged turbulence model with the scheme based on FVM by SIMPLE series algorithm was established and applied to spur-dike flow computation. In this model, the over-relaxed approach was adopted to estimate the diffusion flux in view of its advantages in reducing errors and sustaining numerical stability usually encountered in non-orthogonal meshes. Two spur-dike cases with different defection angles (90oand 135o) were analyzed to validate the model. Computed results show that the predicted velocities and recirculation lengths are in good agreement with the observed data. Moreover, the computations on structured and unstructured grids were compared in terms of the approximately equivalent grid numbers. It can be concluded that the precision with unstructured grids is higher than that with structured grids in spite that the CPU time required is slightly more with unstructured grids. Thus, it is significant to apply the method to numerical simulation of practical hydraulic engineering.

  11. Decomposing Scanned Assembly Meshes Based on Periodicity Recognition and Its Application to Kinematic Simulation Modeling

    Science.gov (United States)

    Mizoguchi, Tomohiro; Kanai, Satoshi

    Along with the rapid growth of industrial X-ray CT scanning systems, it is now possible to non-destructively acquire the entire meshes of assemblies consisting of a set of parts. For the advanced inspections of the assemblies, such as estimation of their assembling errors or examinations of their behaviors in the motions, based on their CT scanned meshes, it is necessary to accurately decompose the mesh and to extract a set of partial meshes each of which correspond to a part. Moreover it is required to create models which can be used for the real-product based simulations. In this paper, we focus on CT scanned meshes of gear assemblies as examples and propose beneficial methods for establishing such advance inspections of the assemblies. We first propose a method that accurately decomposes the mesh into partial meshes each of which corresponds to a gear based on periodicity recognitions. The key idea is first to accurately recognize the periodicity of each gear and then to extract the partial meshes as sets of topologically connected mesh elements where periodicities are valid. Our method can robustly and accurately recognize periodicities from noisy scanned meshes. In contrast to previous methods, our method can deal with single-material CT scanned meshes and can estimate the correct boundaries of neighboring parts with no previous knowledge. Moreover it can efficiently extract the partial meshes from large scanned meshes containing about one million triangles in a few minutes. We also propose a method for creating simulation models which can be used for a gear teeth contact evaluation using extracted partial meshes and their periodicities. Such an evaluation of teeth contacts is one of the most important functions in kinematic simulations of gear assemblies for predicting the power transmission efficiency, noise and vibration. We demonstrate the effectiveness of our method on a variety of artificial and CT scanned meshes.

  12. Development and application of a GIS-based sediment budget model.

    Science.gov (United States)

    Ramos-Scharrón, Carlos E; Macdonald, Lee H

    2007-07-01

    Accelerated erosion and increased sediment yields resulting from changes in land use are a critical environmental problem. Resource managers and decision makers need spatially explicit tools to help them predict the changes in sediment production and delivery due to unpaved roads and other types of land disturbance. This is a particularly important issue in much of the Caribbean because of the rapid pace of development and potential damage to nearshore coral reef communities. The specific objectives of this study were to: (1) develop a GIS-based sediment budget model; (2) use the model to evaluate the effects of unpaved roads on sediment delivery rates in three watersheds on St. John in the US Virgin Islands; and (3) compare the predicted sediment yields to pre-existing data. The St. John Erosion Model (STJ-EROS) is an ArcInfo-based program that uses empirical sediment production functions and delivery ratios to quantify watershed-scale sediment yields. The program consists of six input routines and five routines to calculate sediment production and delivery. The input routines have interfaces that allow the user to adjust the key variables that control sediment production and delivery. The other five routines use pre-set erosion rate constants, user-defined variables, and values from nine data layers to calculate watershed-scale sediment yields from unpaved road travelways, road cutslopes, streambanks, treethrow, and undisturbed hillslopes. STJ-EROS was applied to three basins on St. John with varying levels of development. Predicted sediment yields under natural conditions ranged from 2 to 7Mgkm(-2)yr(-1), while yield rates for current conditions ranged from 8 to 46Mgkm(-2)yr(-1). Unpaved roads are estimated to be increasing sediment delivery rates by 3-6 times for Lameshur Bay, 5-9 times for Fish Bay, and 4-8 times for Cinnamon Bay. Predicted basin-scale sediment yields for both undisturbed and current conditions are within the range of measured sediment yields

  13. An overview of current applications, challenges, and future trends in distributed process-based models in hydrology

    Science.gov (United States)

    Fatichi, Simone; Vivoni, Enrique R.; Odgen, Fred L; Ivanov, Valeriy Y; Mirus, Benjamin B.; Gochis, David; Downer, Charles W; Camporese, Matteo; Davison, Jason H; Ebel, Brian A.; Jones, Norm; Kim, Jongho; Mascaro, Giuseppe; Niswonger, Richard; Restrepo, Pedro; Rigon, Riccardo; Shen, Chaopeng; Sulis, Mauro; Tarboton, David

    2016-01-01

    Process-based hydrological models have a long history dating back to the 1960s. Criticized by some as over-parameterized, overly complex, and difficult to use, a more nuanced view is that these tools are necessary in many situations and, in a certain class of problems, they are the most appropriate type of hydrological model. This is especially the case in situations where knowledge of flow paths or distributed state variables and/or preservation of physical constraints is important. Examples of this include: spatiotemporal variability of soil moisture, groundwater flow and runoff generation, sediment and contaminant transport, or when feedbacks among various Earth’s system processes or understanding the impacts of climate non-stationarity are of primary concern. These are situations where process-based models excel and other models are unverifiable. This article presents this pragmatic view in the context of existing literature to justify the approach where applicable and necessary. We review how improvements in data availability, computational resources and algorithms have made detailed hydrological simulations a reality. Avenues for the future of process-based hydrological models are presented suggesting their use as virtual laboratories, for design purposes, and with a powerful treatment of uncertainty.

  14. An overview of current applications, challenges, and future trends in distributed process-based models in hydrology

    Science.gov (United States)

    Fatichi, Simone; Vivoni, Enrique R.; Ogden, Fred L.; Ivanov, Valeriy Y.; Mirus, Benjamin; Gochis, David; Downer, Charles W.; Camporese, Matteo; Davison, Jason H.; Ebel, Brian; Jones, Norm; Kim, Jongho; Mascaro, Giuseppe; Niswonger, Richard; Restrepo, Pedro; Rigon, Riccardo; Shen, Chaopeng; Sulis, Mauro; Tarboton, David

    2016-06-01

    Process-based hydrological models have a long history dating back to the 1960s. Criticized by some as over-parameterized, overly complex, and difficult to use, a more nuanced view is that these tools are necessary in many situations and, in a certain class of problems, they are the most appropriate type of hydrological model. This is especially the case in situations where knowledge of flow paths or distributed state variables and/or preservation of physical constraints is important. Examples of this include: spatiotemporal variability of soil moisture, groundwater flow and runoff generation, sediment and contaminant transport, or when feedbacks among various Earth's system processes or understanding the impacts of climate non-stationarity are of primary concern. These are situations where process-based models excel and other models are unverifiable. This article presents this pragmatic view in the context of existing literature to justify the approach where applicable and necessary. We review how improvements in data availability, computational resources and algorithms have made detailed hydrological simulations a reality. Avenues for the future of process-based hydrological models are presented suggesting their use as virtual laboratories, for design purposes, and with a powerful treatment of uncertainty.

  15. Application of model-based spectral analysis to wind-profiler radar observations

    Energy Technology Data Exchange (ETDEWEB)

    Boyer, E. [ENS, Cachan (France). LESiR; Petitdidier, M.; Corneil, W. [CETP, Velizy (France); Adnet, C. [THALES Air Dfense, Bagneux (France); Larzabal, P. [ENS, Cachan (France). LESiR; IUT, Cachan (France). CRIIP

    2001-08-01

    A classical way to reduce a radar's data is to compute the spectrum using FFT and then to identify the different peak contributions. But in case an overlapping between the different echoes (atmospheric echo, clutter, hydrometer echo..) exists, Fourier-like techniques provide poor frequency resolution and then sophisticated peak-identification may not be able to detect the different echoes. In order to improve the number of reduced data and their quality relative to Fourier spectrum analysis, three different methods are presented in this paper and applied to actual data. Their approach consists of predicting the main frequency-components, which avoids the development of very sophisticated peak-identification algorithms. The first method is based on cepstrum properties generally used to determine the shift between two close identical echoes. We will see in this paper that this method cannot provide a better estimate than Fourier-like techniques in an operational use. The second method consists of an autoregressive estimation of the spectrum. Since the tests were promising, this method was applied to reduce the radar data obtained during two thunderstorms. The autoregressive method, which is very simple to implement, improved the Doppler-frequency data reduction relative to the FFT spectrum analysis. The third method exploits a MUSIC algorithm, one of the numerous subspace-based methods, which is well adapted to estimate spectra composed of pure lines. A statistical study of performances of this method is presented, and points out the very good resolution of this estimator in comparison with Fourier-like techniques. Application to actual data confirms the good qualities of this estimator for reducing radar's data. (orig.)

  16. Application of model-based spectral analysis to wind-profiler radar observations

    Science.gov (United States)

    Boyer, E.; Petitdidier, M.; Corneil, W.; Adnet, C.; Larzabal, P.

    2001-08-01

    A classical way to reduce a radar’s data is to compute the spectrum using FFT and then to identify the different peak contributions. But in case an overlapping between the different echoes (atmospheric echo, clutter, hydrometeor echo. . . ) exists, Fourier-like techniques provide poor frequency resolution and then sophisticated peak-identification may not be able to detect the different echoes. In order to improve the number of reduced data and their quality relative to Fourier spectrum analysis, three different methods are presented in this paper and applied to actual data. Their approach consists of predicting the main frequency-components, which avoids the development of very sophisticated peak-identification algorithms. The first method is based on cepstrum properties generally used to determine the shift between two close identical echoes. We will see in this paper that this method cannot provide a better estimate than Fourier-like techniques in an operational use. The second method consists of an autoregressive estimation of the spectrum. Since the tests were promising, this method was applied to reduce the radar data obtained during two thunder-storms. The autoregressive method, which is very simple to implement, improved the Doppler-frequency data reduction relative to the FFT spectrum analysis. The third method exploits a MUSIC algorithm, one of the numerous subspace-based methods, which is well adapted to estimate spectra composed of pure lines. A statistical study of performances of this method is presented, and points out the very good resolution of this estimator in comparison with Fourier-like techniques. Application to actual data confirms the good qualities of this estimator for reducing radar’s data.

  17. Supply Chain Vulnerability Analysis Using Scenario-Based Input-Output Modeling: Application to Port Operations.

    Science.gov (United States)

    Thekdi, Shital A; Santos, Joost R

    2016-05-01

    Disruptive events such as natural disasters, loss or reduction of resources, work stoppages, and emergent conditions have potential to propagate economic losses across trade networks. In particular, disruptions to the operation of container port activity can be detrimental for international trade and commerce. Risk assessment should anticipate the impact of port operation disruptions with consideration of how priorities change due to uncertain scenarios and guide investments that are effective and feasible for implementation. Priorities for protective measures and continuity of operations planning must consider the economic impact of such disruptions across a variety of scenarios. This article introduces new performance metrics to characterize resiliency in interdependency modeling and also integrates scenario-based methods to measure economic sensitivity to sudden-onset disruptions. The methods will be demonstrated on a U.S. port responsible for handling $36.1 billion of cargo annually. The methods will be useful to port management, private industry supply chain planning, and transportation infrastructure management. PMID:26271771

  18. Viscoelastic model based force control for soft tissue interaction and its application in physiological motion compensation.

    Science.gov (United States)

    Moreira, Pedro; Zemiti, Nabil; Liu, Chao; Poignet, Philippe

    2014-09-01

    Controlling the interaction between robots and living soft tissues has become an important issue as the number of robotic systems inside the operating room increases. Many researches have been done on force control to help surgeons during medical procedures, such as physiological motion compensation and tele-operation systems with haptic feedback. In order to increase the performance of such controllers, this work presents a novel force control scheme using Active Observer (AOB) based on a viscoelastic interaction model. The control scheme has shown to be stable through theoretical analysis and its performance was evaluated by in vitro experiments. In order to evaluate how the force control scheme behaves under the presence of physiological motion, experiments considering breathing and beating heart disturbances are presented. The proposed control scheme presented a stable behavior in both static and moving environment. The viscoelastic AOB presented a compensation ratio of 87% for the breathing motion and 79% for the beating heart motion. PMID:24612709

  19. Large eddy simulation based fire modeling applications for Indian nuclear power plant

    International Nuclear Information System (INIS)

    Full text of publication follows: The Nuclear Power Plants (NPPs) are always designed for the highest level of safety against postulated accidents which may be initiated due to internal or external causes. One of the external/internal causes, which may lead to accident in the reactor and its associated systems, is fire in certain vital areas of the plant. Conventionally, the fire containment approach and/or the fire confinement approach is used in designing the fire protection systems of NPPs. Indian NPPs (PHWRs) follow the combined approach to ensure plant safety and all newly designed plants are required to comply with the provisions of Atomic Energy Regulatory Board (AERB) fire safety Guide. In respect of older plants, the reassessment of adequacy of fire safety provisions in the light of current advances has becomes essential so as to decide upon the steps for retrofitting. Keeping this in mind the deterministic fire hazard analysis was carried out for the Madras Atomic Power Station (MAPS). As a part of this exercise, detailed fire consequences analysis was required to be carried out for various critical areas. The choice of CFD based code was considered appropriate for these studies. A dedicated fire hazard analysis code Fire Dynamics Simulator (FDS) from NIST was used to perform these case studies. The code has option to use advanced fire models based on Large Eddy Simulation (LES) technique/ Direct Numerical Simulation (DNS) to model the fire-generated conditions. The LES option has been extensively used in the present studies which were primarily aimed at estimating the damage time for important safety related cable. Present paper describes the salient features of the methodology and important results for one of the most critical areas i.e. cable bridge area of MAPS. The typical dimensions of the cable bridge area are (length x breadth x height) of 12 m x 6 m x 2.5 m with an opening on one side of the cable bridge area. With almost equal gap, six numbers

  20. Model-Based Testing for Web Applications%基于模型的Web应用测试

    Institute of Scientific and Technical Information of China (English)

    缪淮扣; 陈圣波; 曾红卫

    2011-01-01

    提出了基于模型的Web应用测试方法,包括建模、测试用例生成、测试用例的执行、模型以及测试用例的可视化等关键技术.设计并实现一个基于模型的Web应用测试系统.以FSM作为被测Web应用的形式测试模型,集成了模型转换器、测试目标分析器、测试序列生成器、FSM和测试序列可视化以及Web应用测试执行引擎等工具.除支持状态覆盖、迁移覆盖、迁移对覆盖等传统的覆盖准则外,还改进/提出了优化状态迁移覆盖、完整消息传递覆盖、完整功能交互覆盖和功能循环交互覆盖等覆盖准则.该文以兴宁水库移民信息管理系统为例演示了该系统.%In this paper, a testing approach to model-based testing for Web applications is proposed which involves in Web application modeling, test generation, test execution and visualization for Web models and test sequences. The authors design and implement a model-based testing system for Web applications while the FSM is regarded as a formal testing model of Web applications under test. And this system integrates Model Transformer, Test Purposes Analyzer, Test Sequences Generator, Visualization tools for FSM and test sequences, Test Execution Engine,etc. Furthermore, the system not only supports the traditional test coverage criteria such as State Coverage, Transition Coverage, Transition Pair Coverage, but also the criteria proposed and improved including Optimized State and Transition Coverage, Complete Message Pass Coverage,Complete Function Interaction Coverage and Function Loop Interaction Coverage. Finally, the authors demonstrate the system taking the Xingning Reservoir Resettlement MIS as our Web application under test.

  1. A novel Q-based online model updating strategy and its application in statistical process control for rubber mixing

    Institute of Scientific and Technical Information of China (English)

    Chunying Zhang; Sun Chen; Fang Wu; Kai Song

    2015-01-01

    To overcome the large time-delay in measuring the hardness of mixed rubber, rheological parameters were used to predict the hardness. A novel Q-based model updating strategy was proposed as a universal platform to track time-varying properties. Using a few selected support samples to update the model, the strategy could dramat-ical y save the storage cost and overcome the adverse influence of low signal-to-noise ratio samples. Moreover, it could be applied to any statistical process monitoring system without drastic changes to them, which is practical for industrial practices. As examples, the Q-based strategy was integrated with three popular algorithms (partial least squares (PLS), recursive PLS (RPLS), and kernel PLS (KPLS)) to form novel regression ones, QPLS, QRPLS and QKPLS, respectively. The applications for predicting mixed rubber hardness on a large-scale tire plant in east China prove the theoretical considerations.

  2. Modeling and Application of a Rapid Fluorescence-Based Assay for Biotoxicity in Anaerobic Digestion.

    Science.gov (United States)

    Chen, Jian Lin; Steele, Terry W J; Stuckey, David C

    2015-11-17

    The sensitivity of anaerobic digestion metabolism to a wide range of solutes makes it important to be able to monitor toxicants in the feed to anaerobic digesters to optimize their operation. In this study, a rapid fluorescence measurement technique based on resazurin reduction using a microplate reader was developed and applied for the detection of toxicants and/or inhibitors to digesters. A kinetic model was developed to describe the process of resazurin reduced to resorufin, and eventually to dihydroresorufin under anaerobic conditions. By modeling the assay results of resazurin (0.05, 0.1, 0.2, and 0.4 mM) reduction by a pure facultative anaerobic strain, Enterococcus faecalis, and fresh mixed anaerobic sludge, with or without 10 mg L(-1) spiked pentachlorophenol (PCP), we found it was clear that the pseudo-first-order rate constant for the reduction of resazurin to resorufin, k1, was a good measure of "toxicity". With lower biomass density and the optimal resazurin addition (0.1 mM), the toxicity of 10 mg L(-1) PCP for E. faecalis and fresh anaerobic sludge was detected in 10 min. By using this model, the toxicity differences among seven chlorophenols to E. faecalis and fresh mixed anaerobic sludge were elucidated within 30 min. The toxicity differences determined by this assay were comparable to toxicity sequences of various chlorophenols reported in the literature. These results suggest that the assay developed in this study not only can quickly detect toxicants for anaerobic digestion but also can efficiently detect the toxicity differences among a variety of similar toxicants. PMID:26457928

  3. MCCI analysis based on the 'segregation phase model', with application in the TOLBIAC-ICB code

    International Nuclear Information System (INIS)

    Full text of publication follows: In the event of a severe accident in a Pressurized Water Reactor, corium, a mixture of molten materials issued from the fuel, cladding and structural elements, appears in the reactor core. One of the scenario of the severe accidents assumes that corium melts through the reactor pressure vessel and spreads over the concrete basemat of the reactor pit. The main question that has to be addressed in these scenario is whether and when the corium will make one's way through the basemat since it would induce groundwater contamination for example. The general approach used in this work is based on the 'segregation phase' model developed by CEA. The solid phase is located at the corium pool boundaries as a solid crust composed of refractory oxides, whereas the corium pool contains no solid. The interfacial temperature between the crust and the pool is the liquidus temperature calculated with the composition of the pool. Thermal-hydraulics (mass and energy balances) is then coupled with physico-chemistry (liquidus temperature, crust composition, chemical reactions). The TOLBIAC-ICB code is developed in the frame of an agreement with EDF, in order to simulate MCCI with the phase segregation model. It is coupled with the GEMINI code for the determination of the physico-chemistry variables. The main purpose of this paper is to present the modelling used in TOLBIAC-ICB and some validation calculations using the data of prototypic experiments available in the literature. Part of the attention focuses on material effects highlighted in some tests and reproduced in the numerical simulations. (authors)

  4. Potential application of FoldX force field based protein modeling in zinc finger nucleases design.

    Science.gov (United States)

    He, ZuYong; Mei, Gui; Zhao, ChunPeng; Chen, YaoSheng

    2011-05-01

    Engineered sequence-specific zinc finger nucleases (ZFNs) make the highly efficient modification of eukaryotic genomes possible. However, most current strategies for developing zinc finger nucleases with customized sequence specificities require the construction of numerous tandem arrays of zinc finger proteins (ZFPs), and subsequent largescale in vitro validation of their DNA binding affinities and specificities via bacterial selection. The labor and expertise required in this complex process limits the broad adoption of ZFN technology. An effective computational assisted design strategy will lower the complexity of the production of a pair of functional ZFNs. Here we used the FoldX force field to build 3D models of 420 ZFP-DNA complexes based on zinc finger arrays developed by the Zinc Finger Consortium using OPEN (oligomerized pool engineering). Using nonlinear and linear regression analysis, we found that the calculated protein-DNA binding energy in a modeled ZFP-DNA complex strongly correlates to the failure rate of the zinc finger array to show significant ZFN activity in human cells. In our models, less than 5% of the three-finger arrays with calculated protein-DNA binding energies lower than -13.132 kcal mol(-1) fail to form active ZFNs in human cells. By contrast, for arrays with calculated protein-DNA binding energies higher than -5 kcal mol(-1), as many as 40% lacked ZFN activity in human cells. Therefore, we suggest that the FoldX force field can be useful in reducing the failure rate and increasing efficiency in the design of ZFNs. PMID:21455692

  5. Application of Finite Element Modeling Methods in Magnetic Resonance Imaging-Based Research and Clinical Management

    Science.gov (United States)

    Fwu, Peter Tramyeon

    The medical image is very complex by its nature. Modeling built upon the medical image is challenging due to the lack of analytical solution. Finite element method (FEM) is a numerical technique which can be used to solve the partial differential equations. It utilized the transformation from a continuous domain into solvable discrete sub-domains. In three-dimensional space, FEM has the capability dealing with complicated structure and heterogeneous interior. That makes FEM an ideal tool to approach the medical-image based modeling problems. In this study, I will address the three modeling in (1) photon transport inside the human breast by implanting the radiative transfer equation to simulate the diffuse optical spectroscopy imaging (DOSI) in order to measurement the percent density (PD), which has been proven as a cancer risk factor in mammography. Our goal is to use MRI as the ground truth to optimize the DOSI scanning protocol to get a consistent measurement of PD. Our result shows DOSI measurement is position and depth dependent and proper scanning scheme and body configuration are needed; (2) heat flow in the prostate by implementing the Penne's bioheat equation to evaluate the cooling performance of regional hypothermia during the robot assisted radical prostatectomy for the individual patient in order to achieve the optimal cooling setting. Four factors are taken into account during the simulation: blood abundance, artery perfusion, cooling balloon temperature, and the anatomical distance. The result shows that blood abundance, prostate size, and anatomical distance are significant factors to the equilibrium temperature of neurovascular bundle; (3) shape analysis in hippocampus by using the radial distance mapping, and two registration methods to find the correlation between sub-regional change to the age and cognition performance, which might not reveal in the volumetric analysis. The result gives a fundamental knowledge of normal distribution in young

  6. Electron-phonon coupling in Ni-based binary alloys with application to displacement cascade modeling.

    Science.gov (United States)

    Samolyuk, G D; Béland, L K; Stocks, G M; Stoller, R E

    2016-05-01

    Energy transfer between lattice atoms and electrons is an important channel of energy dissipation during displacement cascade evolution in irradiated materials. On the assumption of small atomic displacements, the intensity of this transfer is controlled by the strength of electron-phonon (el-ph) coupling. The el-ph coupling in concentrated Ni-based alloys was calculated using electronic structure results obtained within the coherent potential approximation. It was found that Ni0.5Fe0.5, Ni0.5Co0.5 and Ni0.5Pd0.5 are ordered ferromagnetically, whereas Ni0.5Cr0.5 is nonmagnetic. Since the magnetism in these alloys has a Stoner-type origin, the magnetic ordering is accompanied by a decrease of electronic density of states at the Fermi level, which in turn reduces the el-ph coupling. Thus, the el-ph coupling values for all alloys are approximately 50% smaller in the magnetic state than for the same alloy in a nonmagnetic state. As the temperature increases, the calculated coupling initially increases. After passing the Curie temperature, the coupling decreases. The rate of decrease is controlled by the shape of the density of states above the Fermi level. Introducing a two-temperature model based on these parameters in 10 keV molecular dynamics cascade simulation increases defect production by 10-20% in the alloys under consideration. PMID:27033732

  7. Electron–phonon coupling in Ni-based binary alloys with application to displacement cascade modeling

    International Nuclear Information System (INIS)

    Energy transfer between lattice atoms and electrons is an important channel of energy dissipation during displacement cascade evolution in irradiated materials. On the assumption of small atomic displacements, the intensity of this transfer is controlled by the strength of electron–phonon (el–ph) coupling. The el–ph coupling in concentrated Ni-based alloys was calculated using electronic structure results obtained within the coherent potential approximation. It was found that Ni0.5Fe0.5, Ni0.5Co0.5 and Ni0.5Pd0.5 are ordered ferromagnetically, whereas Ni0.5Cr0.5 is nonmagnetic. Since the magnetism in these alloys has a Stoner-type origin, the magnetic ordering is accompanied by a decrease of electronic density of states at the Fermi level, which in turn reduces the el–ph coupling. Thus, the el–ph coupling values for all alloys are approximately 50% smaller in the magnetic state than for the same alloy in a nonmagnetic state. As the temperature increases, the calculated coupling initially increases. After passing the Curie temperature, the coupling decreases. The rate of decrease is controlled by the shape of the density of states above the Fermi level. Introducing a two-temperature model based on these parameters in 10 keV molecular dynamics cascade simulation increases defect production by 10–20% in the alloys under consideration. (paper)

  8. Electron–phonon coupling in Ni-based binary alloys with application to displacement cascade modeling

    Science.gov (United States)

    Samolyuk, G. D.; Béland, L. K.; Stocks, G. M.; Stoller, R. E.

    2016-05-01

    Energy transfer between lattice atoms and electrons is an important channel of energy dissipation during displacement cascade evolution in irradiated materials. On the assumption of small atomic displacements, the intensity of this transfer is controlled by the strength of electron–phonon (el–ph) coupling. The el–ph coupling in concentrated Ni-based alloys was calculated using electronic structure results obtained within the coherent potential approximation. It was found that Ni0.5Fe0.5, Ni0.5Co0.5 and Ni0.5Pd0.5 are ordered ferromagnetically, whereas Ni0.5Cr0.5 is nonmagnetic. Since the magnetism in these alloys has a Stoner-type origin, the magnetic ordering is accompanied by a decrease of electronic density of states at the Fermi level, which in turn reduces the el–ph coupling. Thus, the el–ph coupling values for all alloys are approximately 50% smaller in the magnetic state than for the same alloy in a nonmagnetic state. As the temperature increases, the calculated coupling initially increases. After passing the Curie temperature, the coupling decreases. The rate of decrease is controlled by the shape of the density of states above the Fermi level. Introducing a two-temperature model based on these parameters in 10 keV molecular dynamics cascade simulation increases defect production by 10–20% in the alloys under consideration.

  9. Application of mathematical modeling-based algorithms to 'off-carrier' cobalt-60 irradiation processes

    International Nuclear Information System (INIS)

    The irradiation of materials and products 'off carrier' has historically been performed using a 'drop-and-read' methodology whereby the radioisotope source is raised and lowered repeatedly until the desired absorbed dose is achieved. This approach is time consuming from both a manpower and process perspective. Static irradiation-based processes can also be costly because of the need for repeated experimental verification of target dose delivery. In our paper we address the methods used for predicting Ethicon Endo Surgery's (EES's) off-carrier absorbed dose distributions. The scenarios described herein are complex due to the fact that the on-carrier process stream exhibits a wide range of densities and dose rates. The levels of observed complexity are attributed to the 'just-in-time' production strategy and its related requirements as they apply to the programming of EES's cobalt-60 irradiators. Validation of off-carrier processing methodologies requires sophisticated parametric-based systems utilizing mathematical algorithms that predict off-carrier absorbed dose rate relative to the on-carrier process stream components. Irradiation process simulation is achieved using a point kernel computer modeling approach, coupled with database generation and maintenance. Dose prediction capabilities are validated via routine and transfer standard dosimetry

  10. Modeling Evidence-Based Application: Using Team-Based Learning to Increase Higher Order Thinking in Nursing Research

    Directory of Open Access Journals (Sweden)

    Bridget Moore

    2015-06-01

    Full Text Available Nursing practice is comprised of knowledge, theory, and research [1]. Because of its impact on the profession, the appraisal of research evidence is critically important. Future nursing professionals must be introduced to the purpose and utility of nursing research, as early exposure provides an opportunity to embed evidence-based practice (EBP into clinical experiences. The AACN requires baccalaureate education to include an understanding of the research process to integrate reliable evidence to inform practice and enhance clinical judgments [1]. Although the importance of these knowledge competencies are evident to healthcare administrators and nursing leaders within the field, undergraduate students at the institution under study sometimes have difficulty understanding the relevance of nursing research to the baccalaureate prepared nurse, and struggle to grasp advanced concepts of qualitative and quantitative research design and methodologies. As undergraduate nursing students generally have not demonstrated an understanding of the relationship between theoretical concepts found within the undergraduate nursing curriculum and the practical application of these concepts in the clinical setting, the research team decided to adopt an effective pedagogical active learning strategy, team-based learning (TBL. Team-based learning shifts the traditional course design to focus on higher thinking skills to integrate desired knowledge [2]. The purpose of this paper is to discuss the impact of course design with the integration of TBL in an undergraduate nursing research course on increasing higher order thinking. [1] American Association of Colleges of Nursing, The Essentials of Baccalaureate Education for Professional Nursing Practice, Washington, DC: American Association of Colleges of Nursing, 2008. [2] B. Bloom, Taxonomy of Educational Objectives, Handbook I: Cognitive Domain, New York: McKay, 1956.

  11. Approximation of skewed interfaces with tensor-based model reduction procedures: Application to the reduced basis hierarchical model reduction approach

    Science.gov (United States)

    Ohlberger, Mario; Smetana, Kathrin

    2016-09-01

    In this article we introduce a procedure, which allows to recover the potentially very good approximation properties of tensor-based model reduction procedures for the solution of partial differential equations in the presence of interfaces or strong gradients in the solution which are skewed with respect to the coordinate axes. The two key ideas are the location of the interface either by solving a lower-dimensional partial differential equation or by using data functions and the subsequent removal of the interface of the solution by choosing the determined interface as the lifting function of the Dirichlet boundary conditions. We demonstrate in numerical experiments for linear elliptic equations and the reduced basis-hierarchical model reduction approach that the proposed procedure locates the interface well and yields a significantly improved convergence behavior even in the case when we only consider an approximation of the interface.

  12. A Codon-Based Model of Host-Specific Selection in Parasites, with an Application to the Influenza A Virus

    DEFF Research Database (Denmark)

    Forsberg, Ronald; Christiansen, Freddy Bugge

    2003-01-01

    Parasites sometimes expand their host range by acquiring a new host species. Following a host change event, the selective regime acting on a given parasite gene may change due to host-specific adaptive alterations of protein functionality or host-specific immune-mediated selection. We present a...... codon-based model that attempts to include these effects by allowing the position-specific substitution process to change in conjunction with a host change event. Following maximum-likelihood parameter estimation, we employ an empirical Bayesian procedure to identify candidate sites, potentially...... involved in hostspecific adaptation. We discuss the applicability of the model to the more general problem of ascertaining whether the selective regime differs between two groups of related organisms. The utility of the model is illustrated on a dataset of nucleoprotein sequences from the influenza A virus...

  13. Pareto-efficient deployment synthesis for safety-critical applications in seamless model-based development

    OpenAIRE

    Zverlov, Sergey; Khalil, Maged; Chaudhary, Mayank

    2016-01-01

    International audience Increasingly complex functionality in automotive applications demand more and more computing power. As room for computing units in modern vehicles dwindles, centralized ar-chitectures-with larger, more powerful processing units-are the trend. With this trend, applications no longer run on dedicated hardware, but share the same computing resources with others on the centralized platform. Ascertaining efficient deployment and scheduling for co-located applications is c...

  14. A novel physical eco-hydrological model concept for preferential flow based on experimental applications.

    Science.gov (United States)

    Jackisch, Conrad; van Schaik, Loes; Graeff, Thomas; Zehe, Erwin

    2014-05-01

    Preferential flow through macropores often determines hydrological characteristics - especially regarding runoff generation and fast transport of solutes. Macropore settings may yet be very different in nature and dynamics, depending on their origin. While biogenic structures follow activity cycles (e.g. earth worms) and population conditions (e.g. roots), pedogenic and geogenic structures may depend on water stress (e.g. cracks) or large events (e.g. flushed voids between skeleton and soil pipes) or simply persist (e.g. bedrock interface). On the one hand, such dynamic site characteristics can be observed in seasonal changes in its reaction to precipitation. On the other hand, sprinkling experiments accompanied by tracers or time-lapse 3D Ground-Penetrating-Radar are suitable tools to determine infiltration patterns and macropore configuration. However, model representation of the macropore-matrix system is still problematic, because models either rely on effective parameters (assuming well-mixed state) or on explicit advection strongly simplifying or neglecting interaction with the diffusive flow domain. Motivated by the dynamic nature of macropores, we present a novel model approach for interacting diffusive and advective water, solutes and energy transport in structured soils. It solely relies on scale- and process-aware observables. A representative set of macropores (data from sprinkling experiments) determines the process model scale through 1D advective domains. These are connected to a 2D matrix domain which is defined by pedo-physical retention properties. Water is represented as particles. Diffusive flow is governed by a 2D random walk of these particles while advection may take place in the macropore domain. Macropore-matrix interaction is computed as dissipation of the advective momentum of a particle by its experienced drag from the matrix domain. Through a representation of matrix and macropores as connected diffusive and advective domains for water

  15. A three-dimensional slope stability model based on GRASS GIS and its application to the Collazzone area, Central Italy

    Science.gov (United States)

    Mergili, M.; Marchesini, I.; Fellin, W.; Rossi, M.; Raia, S.; Guzzetti, F.

    2012-04-01

    Landslide risk depends on landslide hazard, i.e. the probability of occurrence of a slope failure of a given magnitude within a specified period and in a given area. The occurrence probability of slope failures in an area characterized by a set of geo-environmental parameters gives the landslide susceptibility. Statistical and deterministic methods are used to assess landslide susceptibility. Deterministic models based on limit equilibrium techniques are applied for the analysis of particular types of landslides (e.g., shallow soil slips, debris flows, rock falls), or to investigate the effects of specific triggers, i.e., an intense rainfall event or an earthquake. In particular, infinite slope stability models are used to calculate the spatial probability of shallow slope failures. In these models, the factor of safety is computed on a pixel basis, assuming a slope-parallel, infinite slip surface. Since shallow slope failures coexist locally with deep-seated landslides, infinite slope stability models fail to describe the complexity of the landslide phenomena. Limit equilibrium models with curved sliding surfaces are geometrically more complex, and their implementation with raster-based GIS is a challenging task. Only few attempts were made to develop GIS-based three-dimensional applications of such methods. We present a preliminary implementation of a GIS-based, three-dimensional slope stability model capable of dealing with deep-seated and shallow rotational slope failures. The model is implemented as a raster module (r.rotstab) in the Open Source GIS package GRASS GIS, and makes use of the three-dimensional sliding surface model proposed by Hovland (1977). Given a DEM and a set of thematic layers of geotechnical and hydraulic parameters, the model tests a large number of randomly determined potential ellipsoidal slip surfaces. In addition to ellipsoidal slip surfaces, truncated ellipsoids are tested, which can occur in the presence of weak layers or hard

  16. Generic Models of Wind Turbine Generators for Advanced Applications in a VSC-based Offshore HVDC Network

    DEFF Research Database (Denmark)

    Zeni, Lorenzo; Margaris, Ioannis; Hansen, Anca Daniela; Sørensen, Poul Ejnar; Kjær, P.C.

    This paper focuses on generic Type 4 wind turbine generators models, their applicability in modern HVDC connections and their capability to provide advanced ancillary services therefrom. A point-to-point HVDC offshore connection is considered. Issues concerning coordinated HVDC and wind farm...... control as well as the need of a communication link are discussed. Two possible control configurations are presented and compared. The first is based on a communication link transmitting the onshore frequency directly to the wind power plant, while the second makes use of a coordinated control scheme...

  17. Modelling Foundations and Applications

    DEFF Research Database (Denmark)

    This book constitutes the refereed proceedings of the 8th European Conference on Modelling Foundations and Applications, held in Kgs. Lyngby, Denmark, in July 2012. The 20 revised full foundations track papers and 10 revised full applications track papers presented were carefully reviewed and sel...

  18. Hyper-hemispheric lens distortion model for 3D-imaging SPAD-array-based applications

    Science.gov (United States)

    Pernechele, Claudio; Villa, Federica A.

    2015-09-01

    Panoramic omnidirectional lenses have the typical draw-back effect to obscure the frontal view, producing the classic "donut-shape" image in the focal plane. We realized a panoramic lens in which the frontal field is make available to be imaged in the focal plane together with the panoramic field, producing a FoV of 360° in azimuth and 270° in elevation; it have then the capabilities of a fish eye plus those of a panoramic lens: we call it hyper-hemispheric lens. We built and test an all-spherical hyper-hemispheric lens. The all-spherical configuration suffer for the typical issues of all ultra wide angle lenses: there is a large distortion at high view angles. The fundamental origin of the optical problems resides on the fact that chief rays angles on the object side are not preserved passing through the optics preceding the aperture stop (fore-optics). This effect produce an image distortion on the focal plane, with the focal length changing along the elevation angles. Moreover, the entrance pupil is shifting at large angle, where the paraxial approximation is not more valid, and tracing the rays appropriately require some effort to the optical designer. It has to be noted here as the distortion is not a source-point-aberrations: it is present also in well corrected optical lenses. Image distortion may be partially corrected using aspheric surface. We describe here how we correct it for our original hyper-hemispheric lens by designing an aspheric surface within the optical train and optimized for a Single Photon Avalanche Diode (SPAD) array-based imaging applications.

  19. Model-Based Systems Engineering for Capturing Mission Architecture System Processes with an Application Case Study - Orion Flight Test 1

    Science.gov (United States)

    Bonanne, Kevin H.

    2011-01-01

    Model-based Systems Engineering (MBSE) is an emerging methodology that can be leveraged to enhance many system development processes. MBSE allows for the centralization of an architecture description that would otherwise be stored in various locations and formats, thus simplifying communication among the project stakeholders, inducing commonality in representation, and expediting report generation. This paper outlines the MBSE approach taken to capture the processes of two different, but related, architectures by employing the Systems Modeling Language (SysML) as a standard for architecture description and the modeling tool MagicDraw. The overarching goal of this study was to demonstrate the effectiveness of MBSE as a means of capturing and designing a mission systems architecture. The first portion of the project focused on capturing the necessary system engineering activities that occur when designing, developing, and deploying a mission systems architecture for a space mission. The second part applies activities from the first to an application problem - the system engineering of the Orion Flight Test 1 (OFT-1) End-to-End Information System (EEIS). By modeling the activities required to create a space mission architecture and then implementing those activities in an application problem, the utility of MBSE as an approach to systems engineering can be demonstrated.

  20. Application of WRF - SWAT OpenMI 2.0 based models integration for real time hydrological modelling and forecasting

    Science.gov (United States)

    Bugaets, Andrey; Gonchukov, Leonid

    2014-05-01

    Intake of deterministic distributed hydrological models into operational water management requires intensive collection and inputting of spatial distributed climatic information in a timely manner that is both time consuming and laborious. The lead time of the data pre-processing stage could be essentially reduced by coupling of hydrological and numerical weather prediction models. This is especially important for the regions such as the South of the Russian Far East where its geographical position combined with a monsoon climate affected by typhoons and extreme heavy rains caused rapid rising of the mountain rivers water level and led to the flash flooding and enormous damage. The objective of this study is development of end-to-end workflow that executes, in a loosely coupled mode, an integrated modeling system comprised of Weather Research and Forecast (WRF) atmospheric model and Soil and Water Assessment Tool (SWAT 2012) hydrological model using OpenMI 2.0 and web-service technologies. Migration SWAT into OpenMI compliant involves reorganization of the model into a separate initialization, performing timestep and finalization functions that can be accessed from outside. To save SWAT normal behavior, the source code was separated from OpenMI-specific implementation into the static library. Modified code was assembled into dynamic library and wrapped into C# class implemented the OpenMI ILinkableComponent interface. Development of WRF OpenMI-compliant component based on the idea of the wrapping web-service clients into a linkable component and seamlessly access to output netCDF files without actual models connection. The weather state variables (precipitation, wind, solar radiation, air temperature and relative humidity) are processed by automatic input selection algorithm to single out the most relevant values used by SWAT model to yield climatic data at the subbasin scale. Spatial interpolation between the WRF regular grid and SWAT subbasins centroid (which are

  1. Model-based engineering of web applications : the flashWeb method

    OpenAIRE

    Jakob, Mihály

    2011-01-01

    Since the first Web page went online in 1990, the rapid development of the World Wide Web has been continuously influencing the way we manage and acquire personal or corporate information. Nowadays, the WWW is much more than a static information medium, which it was at the beginning. The expressive power of modern programming languages is at the full disposal of Web application developers, who continue to surprise Web users with innovative applications having feature sets and complexity, whic...

  2. Geophysical Applications of Vegetation Modeling

    OpenAIRE

    J. O. Kaplan

    2001-01-01

    This thesis describes the development and selected applications of a global vegetation model, BIOME4. The model is applied to problems in high-latitude vegetation distribution and climate, trace gas production, and isotope biogeochemistry. It demonstrates how a modeling approach, based on principles of plant physiology and ecology, can be applied to interdisciplinary problems that cannot be adequately addressed by direct observations or experiments. The work is relevant to understanding the p...

  3. Sznajd Model and its Applications

    International Nuclear Information System (INIS)

    In 2000 we proposed a sociophysics model of opinion formation, which was based on trade union maxim ''United we Stand, Divided we Fall'' (USDF) and latter due to Dietrich Stauffer became known as the Sznajd model (SM). The main difference between SM compared to voter or Ising-type models is that information flows outward. In this paper we review the modifications and applications of SM that have been proposed in the literature. (author)

  4. Sznajd model and its applications

    CERN Document Server

    Sznajd-Weron, K

    2005-01-01

    In 2000 we proposed a sociophysics model of opinion formation, which was based on trade union maxim "United we Stand, Divided we Fall" (USDF) and latter due to Dietrich Stauffer became known as the Sznajd model (SM). The main difference between SM compared to voter or Ising-type models is that information flows outward. In this paper we review the modifications and applications of SM that have been proposed in the literature.

  5. Research of Optical Fiber Coil Winding Model Based on Large-deformation Theory of Elasticity and Its Application

    Institute of Scientific and Technical Information of China (English)

    JIA Ming; YANG Gongliu

    2011-01-01

    Optical fiber coil winding model is used to guide proper and high precision coil winding for fiber optic gyroscope(FOG)application.Based on the large-deformation theory of elasticity,stress analysis of optical fiber free end has been made and balance equation of infinitesimal fiber is deduced,then deformation equation is derived by substituting terminal conditions.On condition that only axial tensile force exists,appmximate curve equation has been built in small angle deformation scope.The comparison of tangent point longitudinal coordinate result between theory and approximation gives constant of integration,and expression with tangent point as origin of coordinate is readjusted.Analyzing the winding parameters of an example,it is clear that the horizontal distance from the highest point of wheel to fiber tangent point has millimeter order of magnitude and significant difference with fiber tension variation,and maintains invariant when wheel radius changes.The height of tensioo and accurate position of tangent point are defined for proper fiber guide.For application to fiber optic gyroscope,spiral-disc winding method and nonideal deformation of straddle section are analyzed,and then spiral-disc quadrupole pattern winding method has been introduced and realized by winding system.The winding results approve that the winding model is applicable.

  6. Application of Slepian theory for improving the accuracy of SH-based global ionosphere models in the Arctic region

    Science.gov (United States)

    Etemadfard, Hossein; Mashhadi Hossainali, Masoud

    2016-03-01

    Due to significant energy resources in polar regions, they have emerged as strategic parts of the world. Consequently, various researches have been funded in order to study these areas in further details. This research intends to improve the accuracy of spherical harmonic (SH)-based Global Ionospheric Models (GIMs) by reconstructing a new map of ionosphere in the Arctic region. For this purpose, the spatiospectral concentration is applied to optimize the base functions. It is carried out using the Slepian theory which was developed by Simons. Here the new base functions and the corresponding coefficients are derived from the SH models for the polar regions. Then, VTEC (vertical total electron content) is reconstructed using Slepian functions and the new coefficients. Reconstructed VTECs and the VTECs derived from SH models are compared to the estimates of this parameter, which are directly derived from dual-frequency GPS measurements. Three International Global Navigation Satellite Systems Service stations located in the northern polar region have been used for this purpose. The starting and ending day of year of adopted GPS data are 69 and 83, respectively, (totally 15 successive days) of the year 2013. According to the obtained results, on average, application of Slepian theory can improve accuracy of the GIM by 1 to 2 total electron content unit (TECU) (1 TECU = 1016 el m-2) in the Arctic region.

  7. Fuzzy Identification Based on T-S Fuzzy Model and Its Application for SCR System

    Science.gov (United States)

    Zeng, Fanchun; Zhang, Bin; Zhang, Lu; Ji, Jinfu; Jin, Wenjing

    An improved T-S model was introduced to identify the model of SCR system. Model structure was selected by physical analyzes and mathematics tests. Three different clustering algorithms were introduced to obtain space partitions. Then, space partitions were amended by mathematics methods. At last, model parameters were identified by least square method. Train data was sampled in 1000MW coal-fired unit SCR system. T-S model of it is identified by three cluster methods. Identify results are proved effective. The merit and demerit among them are analyzed in the end.

  8. Application of Nonlinear Predictive Control Based on RBF Network Predictive Model in MCFC Plant

    Institute of Scientific and Technical Information of China (English)

    CHEN Yue-hua; CAO Guang-yi; ZHU Xin-jian

    2007-01-01

    This paper described a nonlinear model predictive controller for regulating a molten carbonate fuel cell (MCFC). A detailed mechanism model of output voltage of a MCFC was presented at first. However, this model was too complicated to be used in a control system. Consequently, an off line radial basis function (RBF) network was introduced to build a nonlinear predictive model. And then, the optimal control sequences were obtained by applying golden mean method. The models and controller have been realized in the MATLAB environment. Simulation results indicate the proposed algorithm exhibits satisfying control effect even when the current densities vary largely.

  9. Application of statistical emulation to an agent-based model: assortative mating and the reversal of gender inequality in education in Belgium.

    OpenAIRE

    De Mulder, Wim; Grow, André; Molenberghs, Geert; Verbeke, Geert

    2015-01-01

    We describe the application of statistical emulation to the outcomes of an agent-based model. The agent-based model simulates the mechanisms that might have linked the reversal of gender inequality in higher education with observed changes in educational assortative mating in Belgium. Using the statistical emulator as a computationally fast approximation to the expensive agent-based model, it is feasible to use a genetic algorithm in nding the parameter values for which the correspondin...

  10. Application of Structure-Based Models of Mechanical and Thermal Properties on Plasma Sprayed Coatings

    Czech Academy of Sciences Publication Activity Database

    Vilémová, Monika; Matějíček, Jiří; Mušálek, Radek; Nohava, J.

    2012-01-01

    Roč. 21, 3-4 (2012), s. 372-382. ISSN 1059-9630 R&D Projects: GA MŠk ME 901 Institutional research plan: CEZ:AV0Z20430508 Keywords : analytical model * elastic modulus * finite element modeling * image analysis * modeling of properties * thermal conductivity * water stabilized plasma Subject RIV: JK - Corrosion ; Surface Treatment of Materials Impact factor: 1.481, year: 2012 http://www.springerlink.com/content/3m24812367315142/fulltext.pdf

  11. Two-stage robust UC including a novel scenario-based uncertainty model for wind power applications

    International Nuclear Information System (INIS)

    Highlights: • Methodological framework for obtaining Robust Unit Commitment (UC) policies. • Wind-power forecast using a revisited bootstrap predictive inference approach. • Novel scenario-based model for wind-power uncertainty. • Efficient modeling framework for obtaining nearly optimal UC policies in reasonable time. • Effective incorporation of wind-power uncertainty in the UC modeling. - Abstract: The complex processes involved in the determination of the availability of power from renewable energy sources, such as wind power, impose great challenges in the forecasting processes carried out by transmission system operators (TSOs). Nowadays, many of these TSOs use operation planning tools that take into account the uncertainty of the wind-power. However, most of these methods typically require strict assumptions about the probabilistic behavior of the forecast error, and usually ignore the dynamic nature of the forecasting process. In this paper a methodological framework to obtain Robust Unit Commitment (UC) policies is presented; such methodology considers a novel scenario-based uncertainty model for wind power applications. The proposed method is composed by three main phases. The first two phases generate a sound wind-power forecast using a bootstrap predictive inference approach. The third phase corresponds to modeling and solving a one-day ahead Robust UC considering the output of the first phase. The performance of proposed approach is evaluated using as case study a new wind farm to be incorporated into the Northern Interconnected System (NIS) of Chile. A projection of wind-based power installation, as well as different characteristic of the uncertain data, are considered in this study

  12. Sequential application of ligand and structure based modeling approaches to index chemicals for their hH4R antagonism.

    Directory of Open Access Journals (Sweden)

    Matteo Pappalardo

    Full Text Available The human histamine H4 receptor (hH4R, a member of the G-protein coupled receptors (GPCR family, is an increasingly attractive drug target. It plays a key role in many cell pathways and many hH4R ligands are studied for the treatment of several inflammatory, allergic and autoimmune disorders, as well as for analgesic activity. Due to the challenging difficulties in the experimental elucidation of hH4R structure, virtual screening campaigns are normally run on homology based models. However, a wealth of information about the chemical properties of GPCR ligands has also accumulated over the last few years and an appropriate combination of these ligand-based knowledge with structure-based molecular modeling studies emerges as a promising strategy for computer-assisted drug design. Here, two chemoinformatics techniques, the Intelligent Learning Engine (ILE and Iterative Stochastic Elimination (ISE approach, were used to index chemicals for their hH4R bioactivity. An application of the prediction model on external test set composed of more than 160 hH4R antagonists picked from the chEMBL database gave enrichment factor of 16.4. A virtual high throughput screening on ZINC database was carried out, picking ∼ 4000 chemicals highly indexed as H4R antagonists' candidates. Next, a series of 3D models of hH4R were generated by molecular modeling and molecular dynamics simulations performed in fully atomistic lipid membranes. The efficacy of the hH4R 3D models in discrimination between actives and non-actives were checked and the 3D model with the best performance was chosen for further docking studies performed on the focused library. The output of these docking studies was a consensus library of 11 highly active scored drug candidates. Our findings suggest that a sequential combination of ligand-based chemoinformatics approaches with structure-based ones has the potential to improve the success rate in discovering new biologically active GPCR drugs and

  13. An open-source Java-based Toolbox for environmental model evaluation: The MOUSE Software Application

    Science.gov (United States)

    A consequence of environmental model complexity is that the task of understanding how environmental models work and identifying their sensitivities/uncertainties, etc. becomes progressively more difficult. Comprehensive numerical and visual evaluation tools have been developed such as the Monte Carl...

  14. A Bayesian approach to functional-based multilevel modeling of longitudinal data: applications to environmental epidemiology

    Science.gov (United States)

    Berhane, Kiros; Molitor, Nuoo-Ting

    2008-01-01

    Flexible multilevel models are proposed to allow for cluster-specific smooth estimation of growth curves in a mixed-effects modeling format that includes subject-specific random effects on the growth parameters. Attention is then focused on models that examine between-cluster comparisons of the effects of an ecologic covariate of interest (e.g. air pollution) on nonlinear functionals of growth curves (e.g. maximum rate of growth). A Gibbs sampling approach is used to get posterior mean estimates of nonlinear functionals along with their uncertainty estimates. A second-stage ecologic random-effects model is used to examine the association between a covariate of interest (e.g. air pollution) and the nonlinear functionals. A unified estimation procedure is presented along with its computational and theoretical details. The models are motivated by, and illustrated with, lung function and air pollution data from the Southern California Children's Health Study. PMID:18349036

  15. Modelica-based modeling and simulation of a twin screw compressor for heat pump applications

    International Nuclear Information System (INIS)

    A new twin screw compressor has been developed by SRM (Svenska Rotor Maskiner) for use in a new high temperature heat pump using water as refrigerant. This article presents a mathematical model of the thermodynamic process of compression in twin screw compressors. Using a special discretization method, a transient twin screw compressor model has been developed using Modelica in order to study the dry compression cycle of this machine at high temperature levels. The pressure and enthalpy evolution in the control volumes of the model are calculated as a function of the rotational angle of the male rotor using energy and continuity equations. In addition, associated processes encountered in real machines such as variable fluid leakages, water injection and heat losses are modeled and implemented in the main compressor model. A comparison is performed using the model developed, demonstrating the behavior of the compressor and the evolution of its different parameters in different configurations with and without water injection. This comparison shows the need for water injection to avoid compressor failure and improve its efficiency. -- Highlights: • Difficulties related to the compressor limit the development of a high temperature heat pump using water as refrigerant. • A new water vapor double screw compressor has been developed to overcome compression problems. • A dynamic model of this compressor has been developed and simulated using Modelica. • The behavior of the compressor has been identified all along the compression cycle and efficiencies have been calculated

  16. A Bayesian approach to functional-based multilevel modeling of longitudinal data: applications to environmental epidemiology

    OpenAIRE

    Berhane, Kiros; Molitor, Nuoo-Ting

    2008-01-01

    Flexible multilevel models are proposed to allow for cluster-specific smooth estimation of growth curves in a mixed-effects modeling format that includes subject-specific random effects on the growth parameters. Attention is then focused on models that examine between-cluster comparisons of the effects of an ecologic covariate of interest (e.g. air pollution) on nonlinear functionals of growth curves (e.g. maximum rate of growth). A Gibbs sampling approach is used to get posterior mean estima...

  17. Innovative Porous Media Approach in Modeling Biofilm Applications, Human Eye and Nanofluid Based Heat Pipes

    OpenAIRE

    Shafahi, Maryam

    2010-01-01

    Biofilm is a dominant form of existence for bacteria in most natural and synthetic environments. Depending on the application area, they can be useful or harmful. They have a helpful influence in bioremediation, microbial enhanced oil recovery, and metal extraction. On the other hand, biofilms are damaging for water pipes, heat exchangers, submarines and body organs. Formation of biofilm within a porous matrix reduces the pore size and total empty space of the system, altering the porosity an...

  18. AN APPLICATION OF HYBRID CLUSTERING AND NEURAL BASED PREDICTION MODELLING FOR DELINEATION OF MANAGEMENT ZONES

    OpenAIRE

    Babankumar S. Bansod; OP Pandey

    2011-01-01

    Starting from descriptive data on crop yield and various other properties, the aim of this study is to reveal the trends on soil behaviour, such as crop yield. This study has been carried out by developing web application that uses a well known technique- Cluster Analysis. The cluster analysis revealed linkages between soil classes for the same field as well as between different fields, which can be partly assigned to crops rotation and determination of variable soil input rates. A hybrid clu...

  19. Applications for Mission Operations Using Multi-agent Model-based Instructional Systems with Virtual Environments

    Science.gov (United States)

    Clancey, William J.

    2004-01-01

    This viewgraph presentation provides an overview of past and possible future applications for artifical intelligence (AI) in astronaut instruction and training. AI systems have been used in training simulation for the Hubble Space Telescope repair, the International Space Station, and operations simulation for the Mars Exploration Rovers. In the future, robots such as may work as partners with astronauts on missions such as planetary exploration and extravehicular activities.

  20. Distributed Parameter Modelling Applications

    DEFF Research Database (Denmark)

    Sales-Cruz, Mauricio; Cameron, Ian; Gani, Rafiqul

    development of a short-path evaporator. The oil shale processing problem illustrates the interplay amongst particle flows in rotating drums, heat and mass transfer between solid and gas phases. The industrial application considers the dynamics of an Alberta-Taciuk processor, commonly used in shale oil and oil......Here the issue of distributed parameter models is addressed. Spatial variations as well as time are considered important. Several applications for both steady state and dynamic applications are given. These relate to the processing of oil shale, the granulation of industrial fertilizers and the...

  1. Application of Novel Rotation Angular Model for 3D Mouse System Based on MEMS Accelerometers

    Institute of Scientific and Technical Information of China (English)

    QIAN Li; CHEN Wen-yuan; XU Guo-ping

    2009-01-01

    A new scheme is proposed to model 3D angular motion of a revolving regular object with miniature, low-cost micro electro mechanical systems (MEMS) accelerometers (instead of gyroscope), which is employed in 3D mouse system. To sense 3D angular motion, the static property of MEMS accelerometer, sensitive to gravity acceleration, is exploited. With the three outputs of configured accelerometers, the proposed model is implemented to get the rotary motion of the rigid object. In order to validate the effectiveness of the proposed model, an input device is developed with the configuration of the scheme. Experimental results show that a simulated 3D cube can accurately track the rotation of the input device. The result indicates the feasibility and effectiveness of the proposed model in the 3D mouse system.

  2. Quality Model and Artificial Intelligence Base Fuel Ratio Management with Applications to Automotive Engine

    Directory of Open Access Journals (Sweden)

    Mojdeh Piran

    2014-01-01

    Full Text Available In this research, manage the Internal Combustion (IC engine modeling and a multi-input-multi-output artificial intelligence baseline chattering free sliding mode methodology scheme is developed with guaranteed stability to simultaneously control fuel ratios to desired levels under various air flow disturbances by regulating the mass flow rates of engine PFI and DI injection systems. Modeling of an entire IC engine is a very important and complicated process because engines are nonlinear, multi inputs-multi outputs and time variant. One purpose of accurate modeling is to save development costs of real engines and minimizing the risks of damaging an engine when validating controller designs. Nevertheless, developing a small model, for specific controller design purposes, can be done and then validated on a larger, more complicated model. Analytical dynamic nonlinear modeling of internal combustion engine is carried out using elegant Euler-Lagrange method compromising accuracy and complexity. A baseline estimator with varying parameter gain is designed with guaranteed stability to allow implementation of the proposed state feedback sliding mode methodology into a MATLAB simulation environment, where the sliding mode strategy is implemented into a model engine control module (“software”. To estimate the dynamic model of IC engine fuzzy inference engine is applied to baseline sliding mode methodology. The fuzzy inference baseline sliding methodology performance was compared with a well-tuned baseline multi-loop PID controller through MATLAB simulations and showed improvements, where MATLAB simulations were conducted to validate the feasibility of utilizing the developed controller and state estimator for automotive engines. The proposed tracking method is designed to optimally track the desired FR by minimizing the error between the trapped in-cylinder mass and the product of the desired FR and fuel mass over a given time interval.

  3. Application of Levenberg-Marquardt Optimization Algorithm Based Multilayer Neural Networks for Hydrological Time Series Modeling

    OpenAIRE

    Umut Okkan

    2011-01-01

    Recently, Artificial Neural Networks (ANN), which is mathematical modelingtools inspired by the properties of the biological neural system, has been typically used inthe studies of hydrological time series modeling. These modeling studies generally includethe standart feed forward backpropagation (FFBP) algorithms such as gradient-descent,gradient-descent with momentum rate and, conjugate gradient etc. As the standart FFBPalgorithms have some disadvantages relating to the time requirement and...

  4. Rigorous model-based uncertainty quantification with application to terminal ballistics, part I: Systems with controllable inputs and small scatter

    Science.gov (United States)

    Kidane, A.; Lashgari, A.; Li, B.; McKerns, M.; Ortiz, M.; Owhadi, H.; Ravichandran, G.; Stalzer, M.; Sullivan, T. J.

    2012-05-01

    This work is concerned with establishing the feasibility of a data-on-demand (DoD) uncertainty quantification (UQ) protocol based on concentration-of-measure inequalities. Specific aims are to establish the feasibility of the protocol and its basic properties, including the tightness of the predictions afforded by the protocol. The assessment is based on an application to terminal ballistics and a specific system configuration consisting of 6061-T6 aluminum plates struck by spherical S-2 tool steel projectiles at ballistic impact speeds. The system's inputs are the plate thickness and impact velocity and the perforation area is chosen as the sole performance measure of the system. The objective of the UQ analysis is to certify the lethality of the projectile, i.e., that the projectile perforates the plate with high probability over a prespecified range of impact velocities and plate thicknesses. The net outcome of the UQ analysis is an M/U ratio, or confidence factor, of 2.93, indicative of a small probability of no perforation of the plate over its entire operating range. The high-confidence (>99.9%) in the successful operation of the system afforded the analysis and the small number of tests (40) required for the determination of the modeling-error diameter, establishes the feasibility of the DoD UQ protocol as a rigorous yet practical approach for model-based certification of complex systems.

  5. Visual attention modelling and applications. Towards perceptual-based editing methods

    OpenAIRE

    Le Meur, Olivier

    2014-01-01

    This manuscript which constitutes a synthesis document of my research in preparation for my Habilitation degree (Habilitation à Diriger des Recherches) presents the most important outcomes of my research. Since my PhD degree in September 2005, I have been working on two main research themes which are the visual attention and saliency-based image editing. Before delving into the details of my research, a brief presentation of the visual attention and saliency-based image editing is made.Visual...

  6. Reduced order model (ROM) of a pouch type lithium polymer battery based on electrochemical thermal principles for real time applications

    International Nuclear Information System (INIS)

    Accurate and fast estimation of state of charge and health during operations plays pivotal roles in prevention of overcharge or undercharge and accurately monitor the state of cells degradation, which requires a model that can be embedded in the battery management system. Currently, models used are based on either empirical equations or electric equivalent circuit components with voltage sources or a combination of the two. The models are relatively simple, but limited to represent a wide range of operating behaviors that includes the effects of temperature, state of charge (SOC) and degradation, to name a few. On the other hand, full order models (FOM) are multi-dimensional or multi-scale models based on electrochemical and thermal principles capable of representing the details of the cell behavior, but inadequate for real time applications, simply because of the high computational time. Therefore, there are needs for the development of a model with an intermediate performance and real time capability, which is accomplished by reduction of the FOM that is called a reduced order model (ROM). The battery used for development of the ROM is a pouch type of lithium ion polymer battery (LiPB) made of LiMn2O4 (LMO)/carbon. The reduction of the model was carried out for ion concentrations, potentials and kinetics, the ion concentration in the electrode using the polynomial approach and the ion concentration in the electrolyte using the state space method, and potentials and electrochemical kinetics by linearization. In addition, the energy equation is used to calculate the cell temperature, on which the diffusion coefficient and the solid electrolyte interphase (SEI) resistance are dependent. The computational time step is determined based on the total computational time and errors at a given SOC range and different C-rates. ROM responses are compared with those of the FOM and experimental data at a single cycle and multiple cycles under different operating conditions

  7. A real options based model and its application to China's overseas oil investment decisions

    International Nuclear Information System (INIS)

    This paper applies real options theory to overseas oil investment by adding an investment-environment factor to oil-resource valuation. A real options model is developed to illustrate how an investor country (or oil company) can evaluate and compare the critical value of oil-resource investment in different countries under oil-price, exchange-rate, and investment-environment uncertainties. The aim is to establish a broad model that can be used by every oil investor country to value overseas oil resources. The model developed here can match three key elements: 1) deal with overseas investment (the effects of investment environment and exchange rates); 2) deal with oil investment (oil price, production decline rate and development cost etc.); 3) the comparability of the results from different countries (different countries' oil-investment situation can be compared by using the option value index (OVI)). China's overseas oil investment is taken as an example to explain the model by calculating each oil-investee country's critical value per unit of oil reserves and examining the effect of different factors on the critical value. The results show that the model developed here can provide useful advice for China's overseas oil investment program. The research would probably also be helpful to other investor countries looking to invest in overseas oil resources. (author)

  8. Soft Sensing Modelling Based on Optimal Selection of Secondary Variables and Its Application

    Institute of Scientific and Technical Information of China (English)

    Qi Li; Cheng Shao

    2009-01-01

    The composition of the distillation column is a very important quality value in refineries, unfortunately, few hardware sensors are available on-line to measure the distillation compositions. In this paper, a novel method using sensitivity matrix analysis and kernel ridge regression (KRR) to implement on-line soft sensing of distillation compositions is proposed. In this approach, the sensitivity matrix analysis is presented to select the most suitable secondary variables to be used as the soft sensor's input. The KRR is used to build the composition soft sensor. Application to a simulated distillation column demonstrates the effectiveness of the method.

  9. Review of intermediate energy nuclear reaction models for accelerator-based nuclear energy applications

    International Nuclear Information System (INIS)

    In this contribution, a brief survey of some important nuclear reaction models that can be helpful for a global analysis of intermediate-energy nucleon-induced reactions has been presented. Essentially, there are two energy regions: the intranuclear cascade regime, where classical Monte Carlo methods are sufficient for a proper description of nuclear reactions and, for energies below about 150 MeV, the regime where more different specific approaches are required. Probably, the best overall picture is obtained if these two different approaches are employed as complementary tools in nuclear data evaluation. A more extensive comparison between the various models has been performed in a recent computer benchmark. (orig.)

  10. Life-Stage Physiologically-Based Pharmacokinetic (PBPK) Model Applications to Screen Environmental Hazards.

    Science.gov (United States)

    This presentation discusses methods used to extrapolate from in vitro high-throughput screening (HTS) toxicity data for an endocrine pathway to in vivo for early life stages in humans, and the use of a life stage PBPK model to address rapidly changing physiological parameters. A...

  11. Cloud computing models and their application in LTE based cellular systems

    NARCIS (Netherlands)

    Staring, A.J.; Karagiannis, G.

    2013-01-01

    As cloud computing emerges as the next novel concept in computer science, it becomes clear that the model applied in large data storage systems used to resolve issues coming forth from an increasing demand, could also be used to resolve the very high bandwidth requirements on access network, core ne

  12. Nonlinear modeling, strength-based design, and testing of flexible piezoelectric energy harvesters under large dynamic loads for rotorcraft applications

    Science.gov (United States)

    Leadenham, Stephen; Erturk, Alper

    2014-04-01

    There has been growing interest in enabling wireless health and usage monitoring for rotorcraft applications, such as helicopter rotor systems. Large dynamic loads and acceleration fluctuations available in these environments make the implementation of vibration-based piezoelectric energy harvesters a very promising choice. However, such extreme loads transmitted to the harvester can also be detrimental to piezoelectric laminates and overall system reliability. Particularly flexible resonant cantilever configurations tuned to match the dominant excitation frequency can be subject to very large deformations and failure of brittle piezoelectric laminates due to excessive bending stresses at the root of the harvester. Design of resonant piezoelectric energy harvesters for use in these environments require nonlinear electroelastic dynamic modeling and strength-based analysis to maximize the power output while ensuring that the harvester is still functional. This paper presents a mathematical framework to design and analyze the dynamics of nonlinear flexible piezoelectric energy harvesters under large base acceleration levels. A strength-based limit is imposed to design the piezoelectric energy harvester with a proof mass while accounting for material, geometric, and dissipative nonlinearities, with a focus on two demonstrative case studies having the same linear fundamental resonance frequency but different overhang length and proof mass values. Experiments are conducted at different excitation levels for validation of the nonlinear design approach proposed in this work. The case studies in this work reveal that harvesters exhibiting similar behavior and power generation performance at low excitation levels (e.g. less than 0.1g) can have totally different strength-imposed performance limitations under high excitations (e.g. above 1g). Nonlinear modeling and strength-based design is necessary for such excitation levels especially when using resonant cantilevers with no

  13. Designing Collaborative E-Learning Environments Based upon Semantic Wiki: From Design Models to Application Scenarios

    Science.gov (United States)

    Li, Yanyan; Dong, Mingkai; Huang, Ronghuai

    2011-01-01

    The knowledge society requires life-long learning and flexible learning environment that enables fast, just-in-time and relevant learning, aiding the development of communities of knowledge, linking learners and practitioners with experts. Based upon semantic wiki, a combination of wiki and Semantic Web technology, this paper designs and develops…

  14. Conceptual design and modeling of particle-matter interaction cooling systems for muon based applications

    CERN Document Server

    Stratakis, Diktys; Rogers, Chris T; Alekou, Androula; Pasternak, Jaroslaw

    2014-01-01

    An ionization cooling channel is a tightly spaced lattice containing absorbers for reducing the momentum of the muon beam, rf cavities for restoring the longitudinal momentum, and strong solenoids for focusing. Such a lattice can be an essential feature for fundamental high-energy physics applications. In this paper we design, simulate, and compare four individual cooling schemes that rely on ionization cooling. We establish a scaling characterizing the impact of rf gradient limitations on the overall performance and systematically compare important lattice parameters such as the required magnetic fields and the number of cavities and absorber lengths for each cooling scenario. We discuss approaches for reducing the peak magnetic field inside the rf cavities by either increasing the lattice cell length or adopting a novel bucked-coil configuration. We numerically examine the performance of our proposed channels with two independent codes that fully incorporate all basic particle-matter-interaction physical pr...

  15. Interacting Particle-based Model for Missing Data in Sensor Networks: Foundations and Applications

    OpenAIRE

    Koushanfar, Farinaz; Kiyavash, Negar; Potkonjak, Miodrag

    2006-01-01

    Missing data is unavoidable in sensor networks due to sensor faults, communication malfunctioning and malicious attacks. There is a very little insight in missing data causes and statistical and pattern properties of missing data in collected data streams. To address this problem, we utilize interacting-particle model that takes into account both patterns of missing data at individual sensor data streams as well as the correlation between occurrence of missing data at other sensor data stream...

  16. Data-based fault-tolerant model predictive controller an application to a complex dearomatization process

    OpenAIRE

    Kettunen, Markus

    2010-01-01

    The tightening global competition during the last few decades has been the driving force for the optimisation of industrial plant operations through the use of advanced control methods, such as model predictive control (MPC). As the occurrence of faults in the process measurements and actuators has become more common due to the increase in the complexity of the control systems, the need for fault-tolerant control (FTC) to prevent the degradation of the controller performance, and therefore th...

  17. Application of polymer model for calculation of oxides activity in B2O3 based melts

    International Nuclear Information System (INIS)

    Possibility of using equations of polymer model for calculation of oxide activity in boron silicate systems is shown. Correlation of calculation and experimental values of MnO activity in MnO-B2O3 MnO-B2O3-SiO2 melts testifies to the fact, that boron coordination number with respect to oxygen in these systems is constant and equals three. 6 refs., 1 figs., 1 tab

  18. A statistical model of catheter motion from interventional x-ray images: application to image-based gating

    International Nuclear Information System (INIS)

    The motion and deformation of catheters that lie inside cardiac structures can provide valuable information about the motion of the heart. In this paper we describe the formation of a novel statistical model of the motion of a coronary sinus (CS) catheter based on principal component analysis of tracked electrode locations from standard mono-plane x-ray fluoroscopy images. We demonstrate the application of our model for the purposes of retrospective cardiac and respiratory gating of x-ray fluoroscopy images in normal dose x-ray fluoroscopy images, and demonstrate how a modification of the technique allows application to very low dose scenarios. We validated our method on ten mono-plane imaging sequences comprising a total of 610 frames from ten different patients undergoing radiofrequency ablation for the treatment of atrial fibrillation. For normal dose images we established systole, end-inspiration and end-expiration gating with success rates of 100%, 92.1% and 86.9%, respectively. For very low dose applications, the method was tested on the same ten mono-plane x-ray fluoroscopy sequences without noise and with added noise at signal to noise ratio (SNR) values of √50, √10, √8, √6, √5, √2 and √1 to simulate the image quality of increasingly lower dose x-ray images. The method was able to detect the CS catheter even in the lowest SNR images with median errors not exceeding 2.6 mm per electrode. Furthermore, gating success rates of 100%, 71.4% and 85.7% were achieved at the low SNR value of √2, representing a dose reduction of more than 25 times. Thus, the technique has the potential to extract useful information whilst substantially reducing the radiation exposure. (paper)

  19. A statistical model of catheter motion from interventional x-ray images: application to image-based gating

    Science.gov (United States)

    Panayiotou, M.; King, A. P.; Ma, Y.; Housden, R. J.; Rinaldi, C. A.; Gill, J.; Cooklin, M.; O'Neill, M.; Rhode, K. S.

    2013-11-01

    The motion and deformation of catheters that lie inside cardiac structures can provide valuable information about the motion of the heart. In this paper we describe the formation of a novel statistical model of the motion of a coronary sinus (CS) catheter based on principal component analysis of tracked electrode locations from standard mono-plane x-ray fluoroscopy images. We demonstrate the application of our model for the purposes of retrospective cardiac and respiratory gating of x-ray fluoroscopy images in normal dose x-ray fluoroscopy images, and demonstrate how a modification of the technique allows application to very low dose scenarios. We validated our method on ten mono-plane imaging sequences comprising a total of 610 frames from ten different patients undergoing radiofrequency ablation for the treatment of atrial fibrillation. For normal dose images we established systole, end-inspiration and end-expiration gating with success rates of 100%, 92.1% and 86.9%, respectively. For very low dose applications, the method was tested on the same ten mono-plane x-ray fluoroscopy sequences without noise and with added noise at signal to noise ratio (SNR) values of √50, √10, √8, √6, √5, √2 and √1 to simulate the image quality of increasingly lower dose x-ray images. The method was able to detect the CS catheter even in the lowest SNR images with median errors not exceeding 2.6 mm per electrode. Furthermore, gating success rates of 100%, 71.4% and 85.7% were achieved at the low SNR value of √2, representing a dose reduction of more than 25 times. Thus, the technique has the potential to extract useful information whilst substantially reducing the radiation exposure.

  20. A dynamic force balance model for colloidal expansion and its DLVO-based application.

    Science.gov (United States)

    Liu, Longcheng; Moreno, Luis; Neretnieks, Ivars

    2009-01-20

    A force balance model that describes the dynamic expansion of colloidal bentonite gels/sols is presented. The colloidal particles are assumed to consist of one or several thin sheets with the other dimensions much larger than their thickness. The forces considered include van der Waals force, diffuse double layer force, thermal force giving rise to Brownian motion, gravity, as well as friction force. The model results in an expression resembling the instationary diffusion equation but with an immensely variable diffusivity. This diffusivity is strongly influenced by the concentration of counterions as well as by the particle concentration in the colloid gel/sol. The properties of the model are explored and discussed, exemplified by the upward expansion of an originally highly compacted bentonite tablet in a test tube. Examples are presented for a number of cases with ionic concentrations varying between very dilute waters up to several molar of counterions. The volume fraction of particles ranges from 40% to very dilute sols. PMID:19105788

  1. WAP - based telemedicine applications

    International Nuclear Information System (INIS)

    Telemedicine refers to the utilization of telecommunication technology for medical diagnosis, treatment, and patient care. Its aim is to provide expert-based health care to remote sites through telecommunication and information technologies. The significant advances in technologies have enabled the introduction of a broad range of telemedicine applications, which are supported by computer networks, wireless communication, and information superhighway. For example, some hospitals are using tele-radiology for remote consultation. Such a system includes medical imaging devices networked with computers and databases. Another growing area is patient monitoring, in which sensors are used to acquire biomedical signals, such as electrocardiogram (ECG), blood pressure, and body temperature, from a remote patient, who could be in bed or moving freely. The signals are then relayed to remote systems for viewing and analysis. Telemedicine can be divided into two basic modes of operations: real-time mode, in which the patient data can be accessed remotely in real-time, and store-and-forward mode, in which the acquired data does not have to be accessed immediately. In the recent years, many parties have demonstrated various telemedicine applications based on the Internet and cellular phone as these two fields have been developing rapidly. A current, recognizable trend in telecommunication is the convergence of wireless communication and computer network technologies. This has been reflected in recently developed telemedicine systems. For example, in 1998 J. Reponen, et al. have demonstrated transmission and display of computerized tomography (CT) examinations using a remote portable computer wirelessly connected to a computer network through TCP/IP on a GSM cellular phone. Two years later, they carried out the same tests with a GSM-based wireless personal digital assistant (PDA). The WAP (Wireless Application Protocol) Forum was founded in 1997 to create a global protocol

  2. A model for requirements traceability in an heterogeneous model-based design process. Application to automotive embedded systems

    OpenAIRE

    Dubois, Hubert; Peraldi-Frati, Marie-Agnès; Lakhal, Fadoi

    2010-01-01

    Requirements traceability modeling is a key issue in real-time embedded design process. In such systems, requirements are of different nature (software-related, system-related, functional and non functional) and must be traced through a multi level design flow which integrates multiple and heterogeneous models. Validation and Verification (V&V) activities must be performed on models and on the final product to check if they are matching the initial require-ments. Results of a design and of V&...

  3. Model-Based Statistical Tracking and Decision Making for Collision Avoidance Application

    OpenAIRE

    Karlsson, Rickard; Jansson, Jonas; Gustafsson, Fredrik

    2004-01-01

    A growing research topic within the automotive industry is active safety systems. These systems aim at helping the driver avoid or mitigate the consequences of an accident. In this paper a collision mitigation system that performs late braking is discussed. The brake decision is based on estimates from tracking sensors. We use a Bayesian approach, implementing an extended Kalman filter (EKF) and a particle filter to solve the tracking problem. The two filters are compared for different sensor...

  4. Constraint-based modelling of mixed microbial populations: Application to polyhydroxyalkanoates production

    OpenAIRE

    Pardelha, Filipa Alexandra Guerreiro

    2013-01-01

    The combined use of mixed microbial cultures (MMC) and fermented feedstock as substrate may significantly decrease polyhydroxyalkanoates (PHA) production costs and make them more competitive in relation to conventional petroleum-based polymers. However, there still exists a lack of knowledge at metabolic level that limits the development of strategies to make this process more effective. In this thesis, system biology computational tools were developed and applied to PHA production by MMC fro...

  5. Model Prediction-Based Approach to Fault Tolerant Control with Applications

    OpenAIRE

    Mahmoud, Professor Magdi S.; Khalid, Dr. Haris M.

    2013-01-01

    Abstract— Fault-tolerant control (FTC) is an integral component in industrial processes as it enables the system to continue robust operation under some conditions. In this paper, an FTC scheme is proposed for interconnected systems within an integrated design framework to yield a timely monitoring and detection of fault and reconfiguring the controller according to those faults. The unscented Kalman filter (UKF)-based fault detection and diagnosis system is initially run on the main plant an...

  6. Applicability of the technology acceptance model for widget-based personal learning environments

    OpenAIRE

    Wild, Fridolin; Ullmann, Thomas; Scott, Peter; Rebedea, Traian; Hoisl, Bernhard

    2011-01-01

    This contribution presents results from two exploratory studies on technology acceptance and use of widget-based personal learning environments. Methodologically, the investigation carried out applies the unified theory of acceptance and use of technology (UTAUT). With the help of this instrument, the study assesses expert judgments about intentions to use and actual use of the emerging technology of flexibly arranged combinations of use-case-sized mini learning tools. This study aims to expl...

  7. B-SPLINE-BASED SVM MODEL AND ITS APPLICATIONS TO OIL WATER-FLOODED STATUS IDENTIFICATION

    Institute of Scientific and Technical Information of China (English)

    Shang Fuhua; Zhao Tiejun; Yi Xiongying

    2007-01-01

    A method of B-spline transform for signal feature extraction is developed. With the B-spline,the log-signal space is mapped into the vector space. An efficient algorithm based on Support Vector Machine (SVM) to automatically identify the water-flooded status of oil-saturated stratum is described.The experiments show that this algorithm can improve the performances for the identification and the generalization in the case of a limited set of samples.

  8. An application to pulmonary emphysema classification based on model of texton learning by sparse representation

    Science.gov (United States)

    Zhang, Min; Zhou, Xiangrong; Goshima, Satoshi; Chen, Huayue; Muramatsu, Chisako; Hara, Takeshi; Yokoyama, Ryojiro; Kanematsu, Masayuki; Fujita, Hiroshi

    2012-03-01

    We aim at using a new texton based texture classification method in the classification of pulmonary emphysema in computed tomography (CT) images of the lungs. Different from conventional computer-aided diagnosis (CAD) pulmonary emphysema classification methods, in this paper, firstly, the dictionary of texton is learned via applying sparse representation(SR) to image patches in the training dataset. Then the SR coefficients of the test images over the dictionary are used to construct the histograms for texture presentations. Finally, classification is performed by using a nearest neighbor classifier with a histogram dissimilarity measure as distance. The proposed approach is tested on 3840 annotated regions of interest consisting of normal tissue and mild, moderate and severe pulmonary emphysema of three subtypes. The performance of the proposed system, with an accuracy of about 88%, is comparably higher than state of the art method based on the basic rotation invariant local binary pattern histograms and the texture classification method based on texton learning by k-means, which performs almost the best among other approaches in the literature.

  9. Overview of Emerging Web 2.0-Based Business Models and Web 2.0 Applications in Businesses: An Ecological Perspective

    OpenAIRE

    In Lee

    2011-01-01

    Web 2.0 offers business organizations an array of new ways to interact with customers and partners. Web 2.0 is continuously evolving and offers new business models and support business processes, customer relationship management, and partner relationship management. This study reviews some of the major business applications of Web 2.0, and identifies Web 2.0-based business models. Six emerging Web 2.0-based business models were identified: (1) Broad Online Community, (2) Focused Online Commun...

  10. Application of PPE Model in Land Adaptability Appraisal in Small Basin Based on RAGA

    Institute of Scientific and Technical Information of China (English)

    2005-01-01

    The paper applies multi-dimension to lower technology-projection pursuit evaluation model in the conservation of water and soil discipline domain, optimizes the projection direction using the improved acceleration genetic algorithms, transforms the multi-dimensional data target to lower sub-space, and values soil adaptability of Dongdagou basin in Keshan County by searching the optimal projection direction and the projection function data. The paper provides a new notion and method for the conservation of water and soil in small basin.

  11. Application of an ion-packing model based on defect clusters to zirconia solid solutions. 2

    International Nuclear Information System (INIS)

    This paper reports that lattice parameter data of cubic phases and cube roots of unit-cell volumes of tetragonal phases in homogeneous ZrO2-containing solid solutions were compiled to examine the validity of Vegard's law. Except for ZrO2--CeO2 and ZrO2--UO2 systems, the data for cubic phases were expressed by the equation d = a,X + B, where d, as, X, and b denote the lattice parameter, a constant depending on dopant species, the dopant content, and a constant independent of dopant species, respectively. For tetragonal phases, the cube roots of unit-cell volumes could be fitted by a similar equation except for the data in the ZrO2--MO2 systems (M = Ge and U). The constant as was calculated using an ion-packing model and was independent of the defect cluster models. The calculated as is close to the experimentally observed one, although the former is slightly smaller than the latter in the ZrO2--MOu systems (u = 1 and 1.5). This difference was ascribed to the lack of consideration of the ionic distortions from the ideal sites of the fluorite-type structure

  12. Towards quantum-based modeling of enzymatic reaction pathways: Application to the acetylholinesterase catalysis

    Science.gov (United States)

    Polyakov, Igor V.; Grigorenko, Bella L.; Moskovsky, Alexander A.; Pentkovski, Vladimir M.; Nemukhin, Alexander V.

    2013-01-01

    We apply computational methods aiming to approach a full quantum mechanical treatment of chemical reactions in proteins. A combination of the quantum mechanical - molecular mechanical methodology for geometry optimization and the fragment molecular orbital approach for energy calculations is examined for an example of acetylcholinesterase catalysis. The codes based on the GAMESS(US) package operational on the 'RSC Tornado' computational cluster are applied to determine that the energy of the reaction intermediate upon hydrolysis of acetylcholine is lower than that of the enzyme-substrate complex. This conclusion is consistent with the experiments and it is free from the empirical force field contributions.

  13. Eye gaze in intelligent user interfaces gaze-based analyses, models and applications

    CERN Document Server

    Nakano, Yukiko I; Bader, Thomas

    2013-01-01

    Remarkable progress in eye-tracking technologies opened the way to design novel attention-based intelligent user interfaces, and highlighted the importance of better understanding of eye-gaze in human-computer interaction and human-human communication. For instance, a user's focus of attention is useful in interpreting the user's intentions, their understanding of the conversation, and their attitude towards the conversation. In human face-to-face communication, eye gaze plays an important role in floor management, grounding, and engagement in conversation.Eye Gaze in Intelligent User Interfac

  14. Time-frequency representation based on time-varying autoregressive model with applications to non-stationary rotor vibration analysis

    Indian Academy of Sciences (India)

    Long Zhang; Guoliang Xiong; Hesheng Liu; Huijun Zou; Weizhong Guo

    2010-04-01

    A parametric time-frequency representation is presented based on timevarying autoregressive model (TVAR), followed by applications to non-stationary vibration signal processing. The identification of time-varying model coefficients and the determination of model order, are addressed by means of neural networks and genetic algorithms, respectively. Firstly, a simulated signal which mimic the rotor vibration during run-up stages was processed for a comparative study on TVAR and other non-parametric time-frequency representations such as Short Time Fourier Transform, Continuous Wavelet Transform, Empirical Mode Decomposition, Wigner–Ville Distribution and Choi–Williams Distribution, in terms of their resolutions, accuracy, cross term suppression as well as noise resistance. Secondly, TVAR was applied to analyse non-stationary vibration signals collected from a rotor test rig during run-up stages, with an aim to extract fault symptoms under non-stationary operating conditions. Simulation and experimental results demonstrate that TVAR is an effective solution to non-stationary signal analysis and has strong capability in signal time-frequency feature extraction.

  15. Linking Satellite Remote Sensing Based Environmental Predictors to Disease: AN Application to the Spatiotemporal Modelling of Schistosomiasis in Ghana

    Science.gov (United States)

    Wrable, M.; Liss, A.; Kulinkina, A.; Koch, M.; Biritwum, N. K.; Ofosu, A.; Kosinski, K. C.; Gute, D. M.; Naumova, E. N.

    2016-06-01

    90% of the worldwide schistosomiasis burden falls on sub-Saharan Africa. Control efforts are often based on infrequent, small-scale health surveys, which are expensive and logistically difficult to conduct. Use of satellite imagery to predictively model infectious disease transmission has great potential for public health applications. Transmission of schistosomiasis requires specific environmental conditions to sustain freshwater snails, however has unknown seasonality, and is difficult to study due to a long lag between infection and clinical symptoms. To overcome this, we employed a comprehensive 8-year time-series built from remote sensing feeds. The purely environmental predictor variables: accumulated precipitation, land surface temperature, vegetative growth indices, and climate zones created from a novel climate regionalization technique, were regressed against 8 years of national surveillance data in Ghana. All data were aggregated temporally into monthly observations, and spatially at the level of administrative districts. The result of an initial mixed effects model had 41% explained variance overall. Stratification by climate zone brought the R2 as high as 50% for major zones and as high as 59% for minor zones. This can lead to a predictive risk model used to develop a decision support framework to design treatment schemes and direct scarce resources to areas with the highest risk of infection. This framework can be applied to diseases sensitive to climate or to locations where remote sensing would be better suited than health surveys.

  16. Application of a Collaborative Filtering Recommendation Algorithm Based on Cloud Model in Intrusion Detection

    Directory of Open Access Journals (Sweden)

    Deguang Wang

    2011-02-01

    Full Text Available Intrusion detection is a computer network system that collects information on several key points. and it gets these information from the security audit, monitoring, attack recognition and response aspects, check if there are some the behavior and signs against the network security policy. The classification of data acquisition is a key part of intrusion detection. In this article, we use the data cloud model to classify the invasion, effectively maintaining a continuous data on the qualitative ambiguity of the concept and evaluation phase of the invasion against the use of the coordination level filtering recommendation algorithm greatly improves the intrusion detection system in the face of massive data processing efficiency suspicious intrusion.

  17. Application of T2 Control Charts and Hidden Markov Models in Condition-Based Maintenance at Thermoelectric Power Plants

    Directory of Open Access Journals (Sweden)

    Emilija Kisić

    2015-01-01

    Full Text Available An innovative approach to condition-based maintenance of coal grinding subsystems at thermoelectric power plants is proposed in the paper. Coal mill grinding tables become worn over time and need to be replaced through time-based maintenance, after a certain number of service hours. At times such replacement is necessary earlier or later than prescribed, depending on the quality of the coal and of the grinding table itself. Considerable financial losses are incurred when the entire coal grinding subsystem is shut down and the grinding table found to not actually require replacement. The only way to determine whether replacement is necessary is to shut down and open the entire subsystem for visual inspection. The proposed algorithm supports condition-based maintenance and involves the application of T2 control charts to distinct acoustic signal parameters in the frequency domain and the construction of Hidden Markov Models whose observations are coded samples from the control charts. In the present research, the acoustic signals were collected by coal mill monitoring at the thermoelectric power plant “Kostolac” in Serbia. The proposed approach provides information about the current condition of the grinding table.

  18. Hybrid Microgrid Model based on Solar Photovoltaics with Batteries and Fuel Cells system for intermittent applications

    Science.gov (United States)

    Patterson, Maxx

    Microgrids are a subset of the modern power structure; using distributed generation (DG) to supply power to communities rather than vast regions. The reduced scale mitigates loss allowing the power produced to do more with better control, giving greater security, reliability, and design flexibility. This paper explores the performance and cost viability of a hybrid grid-tied microgrid that utilizes Photovoltaic (PV), batteries, and fuel cell (FC) technology. The concept proposes that each community home is equipped with more PV than is required for normal operation. As the homes are part of a microgrid, excess or unused energy from one home is collected for use elsewhere within the microgrid footprint. The surplus power that would have been discarded becomes a community asset, and is used to run intermittent services. In this paper, the modeled community does not have parking adjacent to each home allowing for the installment of a privately owned slower Level 2 charger, making EV ownership option untenable. A solution is to provide a Level 3 DC Quick Charger (DCQC) as the intermittent service. The addition of batteries and Fuel Cells are meant to increase load leveling, reliability, and instill limited island capability.

  19. Model-based software design

    Science.gov (United States)

    Iscoe, Neil; Liu, Zheng-Yang; Feng, Guohui; Yenne, Britt; Vansickle, Larry; Ballantyne, Michael

    1992-01-01

    Domain-specific knowledge is required to create specifications, generate code, and understand existing systems. Our approach to automating software design is based on instantiating an application domain model with industry-specific knowledge and then using that model to achieve the operational goals of specification elicitation and verification, reverse engineering, and code generation. Although many different specification models can be created from any particular domain model, each specification model is consistent and correct with respect to the domain model.

  20. Application of model based predictive control to a solvent extraction plant

    International Nuclear Information System (INIS)

    British Nuclear Fuels plc. (BNFL) is the most experienced nuclear fuel company in the world, having supplied nuclear fuel cycle services in the UK and overseas for over forty years. BNFL is one of only two companies in the world that is able to offer nuclear fuel manufacture, enrichment, reprocessing and waste management services. In addition to its work for the UK Nuclear Power Programme, BNFL has developed a substantial export business with nuclear power plant operators in Western Europe, Japan and North America, which now accounts for 18% of the annual turnover. BNFL's plants re situated in North West England and Southern Scotland. Nuclear fuel and fuel products are manufactured at Springfields near Preston; uranium enrichment by the centrifuge process is carried out at Capenhurst, near Chester; reprocessing and waste management services are provided at Sellafield, West Cumbria. The Company's headquarters and engineering design facilities are based at Risley, near Warrington. BNFL also owns and operates two (MAGNOX) nuclear power stations-Calder Hall, on the Sellafield site, the Chapelcross, near Dumfries in Southern Scotland

  1. Application potential of Agent Based Simulation and Discrete Event Simulation in Enterprise integration modelling concepts

    Directory of Open Access Journals (Sweden)

    Paul-Eric DOSSOU

    2013-07-01

    Full Text Available Normal 0 21 false false false EN-US JA X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Tabla normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:12.0pt; font-family:Cambria; mso-ascii-font-family:Cambria; mso-ascii-theme-font:minor-latin; mso-hansi-font-family:Cambria; mso-hansi-theme-font:minor-latin; mso-ansi-language:EN-US;} This paper aims to present the dilemma of simulation tool selection. Authors discuss the examples of methodologies of enterprises architectures (CIMOSA and GRAI where agent approach is used to solve planning and managing problems. Actually simulation is widely used and practically only one tool which can enable verification of complex systems. Many companies face the problem, which simulation tool is appropriate to use for verification. Selected tools based on ABS and DES are presented. Some tools combining DES and ABS approaches are described. Authors give some recommendation on selection process.

  2. Application potential of Agent Based Simulation and Discrete Event Simulation in Enterprise integration modelling concepts

    Directory of Open Access Journals (Sweden)

    Pawel PAWLEWSKI

    2012-07-01

    Full Text Available Normal 0 21 false false false EN-US JA X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Tabla normal"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:12.0pt; font-family:Cambria; mso-ascii-font-family:Cambria; mso-ascii-theme-font:minor-latin; mso-hansi-font-family:Cambria; mso-hansi-theme-font:minor-latin; mso-ansi-language:EN-US;} This paper aims to present the dilemma of simulation tool selection. Authors discuss the examples of methodologies of enterprises architectures (CIMOSA and GRAI where agent approach is used to solve planning and managing problems. Actually simulation is widely used and practically only one tool which can enable verification of complex systems. Many companies face the problem, which simulation tool is appropriate to use for verification. Selected tools based on ABS and DES are presented. Some tools combining DES and ABS approaches are described. Authors give some recommendation on selection process.

  3. Modeling Evidence-Based Application: Using Team-Based Learning to Increase Higher Order Thinking in Nursing Research

    OpenAIRE

    Bridget Moore; Jennifer Styron; Kristina Miller

    2015-01-01

    Nursing practice is comprised of knowledge, theory, and research [1]. Because of its impact on the profession, the appraisal of research evidence is critically important. Future nursing professionals must be introduced to the purpose and utility of nursing research, as early exposure provides an opportunity to embed evidence-based practice (EBP) into clinical experiences. The AACN requires baccalaureate education to include an understanding of the research process to integrate reliable eviden...

  4. Principles of models based engineering

    Energy Technology Data Exchange (ETDEWEB)

    Dolin, R.M.; Hefele, J.

    1996-11-01

    This report describes a Models Based Engineering (MBE) philosophy and implementation strategy that has been developed at Los Alamos National Laboratory`s Center for Advanced Engineering Technology. A major theme in this discussion is that models based engineering is an information management technology enabling the development of information driven engineering. Unlike other information management technologies, models based engineering encompasses the breadth of engineering information, from design intent through product definition to consumer application.

  5. Direct synchronous-asynchronous conversion system for hybrid electrical vehicle applications. An energy-based modeling approach

    OpenAIRE

    Muñoz-Aguilar, Raúl S.; Dòria-Cerezo, Arnau; Puleston, Pablo Federico

    2013-01-01

    This paper presents a proposal for a series hybrid electric vehicle propulsion system. This new configuration is based on a wound-rotor synchronous generator (WRSM) and a doubly-fed induction machine (DFIM). The energy-based model of the whole system is obtained taking advantage of the capabilities of the port-based modeling techniques. From the dq port-controlled Hamiltonian description of the WRSM and DFIM, the Hamiltonian model of the proposed Direct Synchronous-Asynchronous Conversion Sys...

  6. PITBUL: a physics-based modeling package for imaging and tracking of airborne targets for HEL applications including active illumination

    Science.gov (United States)

    Van Zandt, Noah R.; McCrae, Jack E.; Fiorino, Steven T.

    2013-05-01

    Aimpoint acquisition and maintenance is critical to high energy laser (HEL) system performance. This study demonstrates the development by the AFIT/CDE of a physics-based modeling package, PITBUL, for tracking airborne targets for HEL applications, including atmospheric and sensor effects and active illumination, which is a focus of this work. High-resolution simulated imagery of the 3D airborne target in-flight as seen from the laser position is generated using the HELSEEM model, and includes solar illumination, laser illumination, and thermal emission. Both CW and pulsed laser illumination are modeled, including the effects of illuminator scintillation, atmospheric backscatter, and speckle, which are treated at a first-principles level. Realistic vertical profiles of molecular and aerosol absorption and scattering, as well as optical turbulence, are generated using AFIT/CDE's Laser Environmental Effects Definition and Reference (LEEDR) model. The spatially and temporally varying effects of turbulence are calculated and applied via a fast-running wave optical method known as light tunneling. Sensor effects, for example blur, sampling, read-out noise, and random photon arrival, are applied to the imagery. Track algorithms, including centroid and Fitts correlation, as a part of a closed loop tracker are applied to the degraded imagery and scored, to provide an estimate of overall system performance. To gauge performance of a laser system against a UAV target, tracking results are presented as a function of signal to noise ratio. Additionally, validation efforts to date involving comparisons between simulated and experimental tracking of UAVs are presented.

  7. Optimization by means of an analytical heat transfer model of a thermal insulation for CSP applications based on radiative shields

    Science.gov (United States)

    Gaetano, A.; Roncolato, J.; Montorfano, D.; Barbato, M. C.; Ambrosetti, G.; Pedretti, A.

    2016-05-01

    The employment of new gaseous heat transfer fluids as air or CO2, which are cheaper and environmentally friendly, is drawing more and more attention within the field of Concentrated Solar Power applications. However, despite the advantages, their use requires receivers with a larger heat transfer area and flow cross section with a consequent greater volume of thermal insulation. Solid thermal insulations currently used present high thermal inertia which is energetically penalizing during the daily transient phases faced by the main plant components (e.g. receivers). With the aim of overcoming this drawback a thermal insulation based on radiative shields is presented in this study. Starting from an initial layout comprising a solid thermal insulation layer, the geometry was optimized avoiding the use of the solid insulation keeping performance and fulfilling the geometrical constraints. An analytical Matlab model was implemented to assess the system thermal behavior in terms of heat loss taking into account conductive, convective and radiative contributions. Accurate 2D Computational Fluid Dynamics (CFD) simulations were run to validate the Matlab model which was then used to select the most promising among three new different designs.

  8. The Web-based Application Server: Combining earth observation with in-situ data and modelling. ISECA Final Report D3.1

    OpenAIRE

    de Kok, J.-L.; DeCorte, L; Peelaerts, W.

    2014-01-01

    The report describes the purpose, architecture and functionalities of the ISECA Web-based Application Server (WAS). This web-based information system combines earth observation and in-situ data with examples of model simulations related to eutrophication for the 2Seas Territorial Waters of the Southern North Sea. Step-by-step instructions on how to use the WAS are included in this report. More background information on the problem of eutrophication and eutrophication modelling is found in IS...

  9. A Model for Water Quality Assessment Based on the Information Entropy and Its Application in the Case of Huiji River

    Institute of Scientific and Technical Information of China (English)

    BingdongZhao; QingliangZhao; JianhuaMa; HuaGuan

    2004-01-01

    Based on the information entropy, a model for water quality assessment is Using this model, the paper gives a case study on the water quality assessment River. The space-time variation law of the water quality is analyzed also in this result indicates that the model possesses some clear mathematic and physical and it is simple, practical and accurate.

  10. Recent Advances in Development and Application of Physiologically-Based Pharmacokinetic (PBPK) Models: a Transition from Academic Curiosity to Regulatory Acceptance

    OpenAIRE

    Jamei, Masoud

    2016-01-01

    There is a renewed surge of interest in applications of physiologically-based pharmacokinetic (PBPK) models by the pharmaceutical industry and regulatory agencies. Developing PBPK models within a systems pharmacology context allows separation of the parameters pertaining to the animal or human body (the system) from that of the drug and the study design which is essential to develop generic drug-independent models used to extrapolate PK/PD properties in various healthy and patient populations...

  11. Methane emissions from floodplains in the Amazon Basin: challenges in developing a process-based model for global applications

    Science.gov (United States)

    Ringeval, B.; Houweling, S.; van Bodegom, P. M.; Spahni, R.; van Beek, R.; Joos, F.; Röckmann, T.

    2014-03-01

    Tropical wetlands are estimated to represent about 50% of the natural wetland methane (CH4) emissions and explain a large fraction of the observed CH4 variability on timescales ranging from glacial-interglacial cycles to the currently observed year-to-year variability. Despite their importance, however, tropical wetlands are poorly represented in global models aiming to predict global CH4 emissions. This publication documents a first step in the development of a process-based model of CH4 emissions from tropical floodplains for global applications. For this purpose, the LPX-Bern Dynamic Global Vegetation Model (LPX hereafter) was slightly modified to represent floodplain hydrology, vegetation and associated CH4 emissions. The extent of tropical floodplains was prescribed using output from the spatially explicit hydrology model PCR-GLOBWB. We introduced new plant functional types (PFTs) that explicitly represent floodplain vegetation. The PFT parameterizations were evaluated against available remote-sensing data sets (GLC2000 land cover and MODIS Net Primary Productivity). Simulated CH4 flux densities were evaluated against field observations and regional flux inventories. Simulated CH4 emissions at Amazon Basin scale were compared to model simulations performed in the WETCHIMP intercomparison project. We found that LPX reproduces the average magnitude of observed net CH4 flux densities for the Amazon Basin. However, the model does not reproduce the variability between sites or between years within a site. Unfortunately, site information is too limited to attest or disprove some model features. At the Amazon Basin scale, our results underline the large uncertainty in the magnitude of wetland CH4 emissions. Sensitivity analyses gave insights into the main drivers of floodplain CH4 emission and their associated uncertainties. In particular, uncertainties in floodplain extent (i.e., difference between GLC2000 and PCR-GLOBWB output) modulate the simulated emissions by a

  12. Stochastic fractal-based models of heterogeneity in subsurface hydrology: Origins, applications, limitations, and future research questions

    Science.gov (United States)

    Molz, Fred J.; Rajaram, Harihar; Lu, Silong

    2004-03-01

    Modern measurement techniques have shown that property distributions in natural porous and fractured media appear highly irregular and nonstationary in a spatial statistical sense. This implies that direct statistical analyses of the property distributions are not appropriate, because the statistical measures developed will be dependent on position and therefore will be nonunique. An alternative, which has been explored to an increasing degree during the past 20 years, is to consider the class of functions known as nonstationary stochastic processes with spatially stationary increments. When such increment distributions are described by probability density functions (PDFs) of the Gaussian, Levy, or gamma class or PDFs that converge to one of these classes under additions, then one is also dealing with a so-called stochastic fractal, the mathematical theory of which was developed during the first half of the last century. The scaling property associated with such fractals is called self-affinity, which is more general that geometric self-similarity. Herein we review the application of Gaussian and Levy stochastic fractals and multifractals in subsurface hydrology, mainly to porosity, hydraulic conductivity, and fracture roughness, along with the characteristics of flow and transport in such fields. Included are the development and application of fractal and multifractal concepts; a review of the measurement techniques, such as the borehole flowmeter and gas minipermeameter, that are motivating the use of fractal-based theories; the idea of a spatial weighting function associated with a measuring instrument; how fractal fields are generated; and descriptions of the topography and aperture distributions of self-affine fractures. In a somewhat different vein the last part of the review deals with fractal- and fragmentation-based descriptions of fracture networks and the implications for transport in such networks. Broad conclusions include the implication that models

  13. Development of an infrared gaseous radiation band model based on NASA SP-3080 for computational fluid dynamic code validation applications

    OpenAIRE

    Nelson, Edward L.

    1992-01-01

    The increased use of infrared imaging as a flow visualization technique and as a validation technique for computational fluid dynamics (CFD) codes has led to an in-depth study of infrared band models. The ability to create fast and accurate images of airframe and plume infrared emissions often depends on the complexity of the band model. An infrared band model code has been created based largely on the band model published in NASA SP-3080, Handbook of Infrared Radiation from Combustion Gases....

  14. Application of observer-based chaotic synchronization and identifiability to original CSK model for secure information transmission

    OpenAIRE

    Garasym, Oleg; Taralova, Ina; Lozi, René

    2015-01-01

    Received (to be inserted by publisher) Modified Lozi system is analyzed as chaotic PRNG and synchronized via observers. The objective of the study is to investigate chaotic-based encryption method that preserves CSK model advantages, but improves the security level. CSK model have been discussed to message encryp-tion because it implies better resistance against noise, but there are many evidences of the model weaknesses. The investigation provides the original CSK model analyses of secure me...

  15. Morse potential-based model for contacting composite rough surfaces: Application to self-assembled monolayer junctions

    Science.gov (United States)

    Sierra-Suarez, Jonatan A.; Majumdar, Shubhaditya; McGaughey, Alan J. H.; Malen, Jonathan A.; Higgs, C. Fred

    2016-04-01

    This work formulates a rough surface contact model that accounts for adhesion through a Morse potential and plasticity through the Kogut-Etsion finite element-based approximation. Compared to the commonly used Lennard-Jones (LJ) potential, the Morse potential provides a more accurate and generalized description for modeling covalent materials and surface interactions. An extension of this contact model to describe composite layered surfaces is presented and implemented to study a self-assembled monolayer (SAM) grown on a gold substrate placed in contact with a second gold substrate. Based on a comparison with prior experimental measurements of the thermal conductance of this SAM junction [Majumdar et al., Nano Lett. 15, 2985-2991 (2015)], the more general Morse potential-based contact model provides a better prediction of the percentage contact area than an equivalent LJ potential-based model.

  16. Determining iron oxide nanoparticle heating efficiency and elucidating local nanoparticle temperature for application in agarose gel-based tumor model.

    Science.gov (United States)

    Shah, Rhythm R; Dombrowsky, Alexander R; Paulson, Abigail L; Johnson, Margaret P; Nikles, David E; Brazel, Christopher S

    2016-11-01

    Magnetic iron oxide nanoparticles (MNPs) have been developed for magnetic fluid hyperthermia (MFH) cancer therapy, where cancer cells are treated through the heat generated by application of a high frequency magnetic field. This heat has also been proposed as a mechanism to trigger release of chemotherapy agents. In each of these cases, MNPs with optimal heating performance can be used to maximize therapeutic effect while minimizing the required dosage of MNPs. In this study, the heating efficiencies (or specific absorption rate, SAR) of two types of MNPs were evaluated experimentally and then predicted from their magnetic properties. MNPs were also incorporated in the core of poly(ethylene glycol-b-caprolactone) micelles, co-localized with rhodamine B fluorescent dye attached to polycaprolactone to monitor local, nanoscale temperatures during magnetic heating. Despite a relatively high SAR produced by these MNPs, no significant temperature rise beyond that observed in the bulk solution was measured by fluorescence in the core of the magnetic micelles. MNPs were also incorporated into a macro-scale agarose gel system that mimicked a tumor targeted by MNPs and surrounded by healthy tissues. The agarose-based tumor models showed that targeted MNPs can reach hyperthermia temperatures inside a tumor with a sufficient MNP concentration, while causing minimal temperature rise in the healthy tissue surrounding the tumor. PMID:27523991

  17. Modeling the water use efficiency of soybean and maize plants under environmental stresses: application of a synthetic model of photosynthesis-transpiration based on stomatal behavior.

    Science.gov (United States)

    Yu, Gui-Rui; Wang, Qiu-Feng; Zhuang, Jie

    2004-03-01

    Understanding the variability of plant WUE and its control mechanism can promote the comprehension to the coupling relationship of water and carbon cycle in terrestrial ecosystem, which is the foundation for developing water-carbon coupling cycle model. In this paper, we made clear the differences of net assimilation rate, transpiration rate, and WUE between the two species by comparing the experiment data of soybean (Glycine max Merr.) and maize (Zea mays L.) plants under water and soil nutrient stresses. WUE of maize was about two and a half times more than that of soybean in the same weather conditions. Enhancement of water stresses led to the marked decrease of Am and Em of two species, but water stresses of some degree could improve WUE, and this effect was more obvious for soybean. WUE of the two species changed with psiL in a second-order curve relation, and the WUE at high fertilization was higher than that at low fertilization, this effect was especially obvious for maize. Moreover, according to the synthetic model of photosynthesis-transpiration based on stomatal behavior (SMPTSB) presented by Yu et al. (2001), the WUE model and its applicability were discussed with the data measured in this experiment. The WUE estimated by means of the model accorded well with the measured values. However, this model underestimated the WUE for maize slightly, thus further improvement on the original model was made in this study. Finally, by discussing some physiological factors controlling Am and WUE, we made clear the physiological explanation for differences of the relative contributions of stomata- and mesophyll processes to control of Am and WUE, and the applicability of WUE model between the two species. Because the requirement to stomatal conductance by unit change of net assimilation rate is different, the responses of opening-closing activity of stomata to environmental stresses are different between the two species. To obtain the same level of net assimilation

  18. PEM Fuel Cells - Fundamentals, Modeling and Applications

    OpenAIRE

    Maher A.R. Sadiq Al-Baghdadi

    2013-01-01

    Part I: Fundamentals Chapter 1: Introduction. Chapter 2: PEM fuel cell thermodynamics, electrochemistry, and performance. Chapter 3: PEM fuel cell components. Chapter 4: PEM fuel cell failure modes. Part II: Modeling and Simulation Chapter 5: PEM fuel cell models based on semi-empirical simulation. Chapter 6: PEM fuel cell models based on computational fluid dynamics. Part III: Applications Chapter 7: PEM fuel cell system design and applications.

  19. PEM Fuel Cells - Fundamentals, Modeling and Applications

    Directory of Open Access Journals (Sweden)

    Maher A.R. Sadiq Al-Baghdadi

    2013-01-01

    Full Text Available Part I: Fundamentals Chapter 1: Introduction. Chapter 2: PEM fuel cell thermodynamics, electrochemistry, and performance. Chapter 3: PEM fuel cell components. Chapter 4: PEM fuel cell failure modes. Part II: Modeling and Simulation Chapter 5: PEM fuel cell models based on semi-empirical simulation. Chapter 6: PEM fuel cell models based on computational fluid dynamics. Part III: Applications Chapter 7: PEM fuel cell system design and applications.

  20. A particle-based model of size or anion exclusion with application to microbial transport in Porous Media

    International Nuclear Information System (INIS)

    This paper presents a novel approach to the quantification of pore-scale exclusion processes, based on the truncation of the distribution of local dispersive displacements in a random-walk particle model. This approach increases the mean velocity of colloidal-sized particles relative to inert solute tracers, and decreases the apparent dispersion. An equivalent continuum (concentration-based) model, with modified velocity and dispersion parameters, is derived. The model was applied to the results of laboratory experiments on bacterial transport in intact cores from a research site near Oyster, Virginia. The significant observed decrease in bacterial arrival times relative to a bromide tracer was closely reproduced by the particle-based model with modest degrees of truncation (8% maximum). The approach provides a conceptually consistent means of incorporating the exclusion process into groundwater transport models

  1. Modelling and simulation of passive Lab-on-a-Chip (LoC) based micromixer for clinical application

    Science.gov (United States)

    Saikat, Chakraborty; Sharath, M.; Srujana, M.; Narayan, K.; Pattnaik, Prasant Kumar

    2016-03-01

    In biomedical application, micromixer is an important component because of many processes requires rapid and efficient mixing. At micro scale, the flow is Laminar due to small channel size which enables controlled rapid mixing. The reduction in analysis time along with high throughput can be achieved with the help of rapid mixing. In LoC application, micromixer is used for mixing of fluids especially for the devices which requires efficient mixing. Micromixer of this type of microfluidic devices with a rapid mixing is useful in application such as DNA/RNA synthesis, drug delivery system & biological agent detection. In this work, we design and simulate a microfluidic based passive rapid micromixer for lab-on-a-chip application.

  2. The effects of composition on glass dissolution rates: The application of four models to a data base

    International Nuclear Information System (INIS)

    Four models have been applied to a data base to relate glass dissolution in distilled water to composition. The data base is used to compare the precisions obtained from the models in fitting actual data. The usefulness of the data base in formulating a model is also demonstrated. Two related models in which the composite or pH-adjusted free energy of hydration of the glass is the correlating parameter are compared with experimental data. In a structural model, the nonbridging oxygen content of the glasses is used to correlate glass dissolution rate to composition. In a model formulated for this report, the cation valence and the oxygen content of the glass are compared with observed dissolution rates. The models were applied to the 28-day normalized silica release at 900C for over 285 glass compositions with surface area to volume ratios of 10 m-1 (Materials Characterization Center MCC-1 glass durability test using distilled water). These glasses included the nonradioactive analogs of WV205 and SRL-165, as well as SRL-131, PNL 76-68, and a European glass, UK209. Predicted glass dissolution rates show similar fits to the data for all four models. The predictions of the models were also plotted for two subsets of the glasses: waste glasses and Savannah River Laboratory glasses. The model predictions fit the data for these groups much better than they fit the data for the entire set of glasses. 14 refs., 12 figs., 7 tabs

  3. Studying nanotube-based oscillators and their application as memory cells via nanoscale continuum modeling and simulation

    OpenAIRE

    Shaoping Xiao; Weixuan Yang

    2015-01-01

    A nanoscale continuum model of carbon nanotube-based oscillators is proposed in this paper. In the continuum model, the nanotube is discretized via the meshfree particle method. The atomistic interlayer interaction between the outer and inner tubes is approximated by the interlayer interaction between particles. The mechanical behaviors of oscillators are studied and compared well with molecular dynamics simulation results. The nanotube-based oscillator can be employed to design a...

  4. Patterns of Use of an Agent-Based Model and a System Dynamics Model: The Application of Patterns of Use and the Impacts on Learning Outcomes

    Science.gov (United States)

    Thompson, Kate; Reimann, Peter

    2010-01-01

    A classification system that was developed for the use of agent-based models was applied to strategies used by school-aged students to interrogate an agent-based model and a system dynamics model. These were compared, and relationships between learning outcomes and the strategies used were also analysed. It was found that the classification system…

  5. Application of thermodynamics-based rate-dependent constitutive models of concrete in the seismic analysis of concrete dams

    Directory of Open Access Journals (Sweden)

    Fei LENG

    2008-09-01

    Full Text Available This paper discusses the seismic analysis of concrete dams with consideration of material nonlinearity. Based on a consistent rate-dependent model and two thermodynamics-based models, two thermodynamics-based rate-dependent constitutive models were developed with consideration of the influence of the strain rate. They can describe the dynamic behavior of concrete and be applied to nonlinear seismic analysis of concrete dams taking into account the rate sensitivity of concrete. With the two models, a nonlinear analysis of the seismic response of the Koyna Gravity Dam and the Dagangshan Arch Dam was conducted. The results were compared with those of a linear elastic model and two rate-independent thermodynamics-based constitutive models, and the influences of constitutive models and strain rate on the seismic response of concrete dams were discussed. It can be concluded from the analysis that, during seismic response, the tensile stress is the control stress in the design and seismic safety evaluation of concrete dams. In different models, the plastic strain and plastic strain rate of concrete dams show a similar distribution. When the influence of the strain rate is considered, the maximum plastic strain and plastic strain rate decrease.

  6. Empirical Differences in Omission Tendency and Reading Ability in PISA: An Application of Tree-Based Item Response Models

    Science.gov (United States)

    Okumura, Taichi

    2014-01-01

    This study examined the empirical differences between the tendency to omit items and reading ability by applying tree-based item response (IRTree) models to the Japanese data of the Programme for International Student Assessment (PISA) held in 2009. For this purpose, existing IRTree models were expanded to contain predictors and to handle…

  7. Application of thermodynamics-based rate-dependent constitutive models of concrete in the seismic analysis of concrete dams

    Institute of Scientific and Technical Information of China (English)

    Leng Fei; Lin Gao

    2008-01-01

    This paper discusses the seismic analysis of concrete dams with consideration of material nonlinearity. Based on a consistent rate-dependent model and two thermodynamics-based models, two thermodynamics-based rate-dependent constitutive models were developed with consideration of the influence of the strain rate. They can describe the dynamic behavior of concrete and be applied to nonlinear seismic analysis of concrete dams taking into account the rate sensitivity of concrete. With the two models, a nonlinear analysis of the seismic response of the Koyna Gravity Dam and the Dagangshan Arch Dam was conducted. The results were compared with those of a linear elastic model and two rate-independent thermodynamics-based constitutive models, and the influences of constitutive models and strain rate on the seismic response of concrete dams were discussed. It can be concluded from the analysis that, during seismic response, the tensile stress is the control stress in the design and seismic safety evaluation of concrete dams. In different models, the plastic strain and plastic strain rate of concrete dams show a similar distribution. When the influence of the strain rate is considered, the maximum plastic strain and plastic strain rate decrease.

  8. A nonparametric urn-based approach to interacting failing systems with an application to credit risk modeling

    CERN Document Server

    Cirillo, Pasquale; Muliere, Pietro

    2010-01-01

    In this paper we propose a new nonparametric approach to interacting failing systems (FS), that is systems whose probability of failure is not negligible in a fixed time horizon, a typical example being firms and financial bonds. The main purpose when studying a FS is to calculate the probability of default and the distribution of the number of failures that may occur during the observation period. A model used to study a failing system is defined default model. In particular, we present a general recursive model constructed by the means of inter- acting urns. After introducing the theoretical model and its properties we show a first application to credit risk modeling, showing how to assess the idiosyncratic probability of default of an obligor and the joint probability of failure of a set of obligors in a portfolio of risks, that are divided into reliability classes.

  9. Model Based Definition

    Science.gov (United States)

    Rowe, Sidney E.

    2010-01-01

    In September 2007, the Engineering Directorate at the Marshall Space Flight Center (MSFC) created the Design System Focus Team (DSFT). MSFC was responsible for the in-house design and development of the Ares 1 Upper Stage and the Engineering Directorate was preparing to deploy a new electronic Configuration Management and Data Management System with the Design Data Management System (DDMS) based upon a Commercial Off The Shelf (COTS) Product Data Management (PDM) System. The DSFT was to establish standardized CAD practices and a new data life cycle for design data. Of special interest here, the design teams were to implement Model Based Definition (MBD) in support of the Upper Stage manufacturing contract. It is noted that this MBD does use partially dimensioned drawings for auxiliary information to the model. The design data lifecycle implemented several new release states to be used prior to formal release that allowed the models to move through a flow of progressive maturity. The DSFT identified some 17 Lessons Learned as outcomes of the standards development, pathfinder deployments and initial application to the Upper Stage design completion. Some of the high value examples are reviewed.

  10. The Garden-Hose Game: A New Model of Computation, and Application to Position-Based Quantum Cryptography

    CERN Document Server

    Buhrman, Harry; Schaffner, Christian; Speelman, Florian

    2011-01-01

    We study position-based cryptography in the quantum setting. We examine a class of protocols that only require the communication of a single qubit and 2n bits of classical information. To this end, we define a new model of communication complexity, the garden-hose model, which enables us to prove upper bounds on the number of EPR pairs needed to attack such schemes. This model furthermore opens up a way to link the security of quantum position-based cryptography to traditional complexity theory.

  11. Roadway management plan based on rockfall modelling calibration and validation. Application along the Ma-10 road in Mallorca (Spain)

    Science.gov (United States)

    Mateos, Rosa Maria; Garcia, Inmaculada; Reichenbach, Paola; Herrera, Gerardo; Sarro, Roberto; Rius, Joan; Aguilo, Raul

    2016-04-01

    The Tramuntana range, in the northwestern sector of the island of Mallorca (Spain), is frequently affected by rockfalls which have caused significant damage, mainly along the road network. The Ma-10 road constitutes the main transportation corridor on the range with a heavy traffic estimated at 7,200 vehicles per day on average. With a length of 111 km and a tortuous path, the road is the connecting track for 12 municipalities and constitutes a strategic road on the island for many tourist resorts. For the period spanning from 1995 to current times, 63 rockfalls have affected the Ma-10 road with volumes ranging from 0.3m3 to 30,000 m3. Fortunately, no fatalities occurred but numerous blockages on the road took place which caused significant economic losses, valued of around 11 MEuro (Mateos el al., 2013). In this work we present the procedure we have applied to calibrate and validate rockfall modelling in the Tramuntana region, using 103 cases of the available detailed rockfall inventory (Mateos, 2006). We have exploited STONE (Guzzetti et al. 2002), a GIS based rockfall simulation software which computes 2D and 3D rockfall trajectories starting from a DTM and maps of the dynamic rolling friction coefficient and of the normal and tangential energy restitution coefficients. The appropriate identification of these parameters determines the accuracy of the simulation. To calibrate them, we have selected 40 rockfalls along the range which include a wide variety of outcropping lithologies. Coefficients values have been changed in numerous attempts in order to select those where the extent and shape of the simulation matched the field mapping. Best results were summarized with the average statistical values for each parameter and for each geotechnical unit, determining that mode values represent more precisely the data. Initially, for the validation stage, 10 well- known rockfalls exploited in the calibration phase have been selected. Confidence tests have been applied

  12. A Model of Application System for Man-Machine-Environment System Engineering in Vessels Based on IDEF0

    Institute of Scientific and Technical Information of China (English)

    Zhen Shang; Changhua Qiu; Shifan Zhu

    2011-01-01

    Applying man-machine-environment system engineering (MMESE) in vessels is a method to improve the effectiveness of the interaction between equipment,environment,and humans for the purpose of advancing operating efficiency,performance,safety,and habitability of a vessel and its subsystems.In the following research,the life cycle of vessels was divided into 9 phases,and 15 research subjects were also identified from among these phases.The 15 subjects were systemized,and then the man-machine-environment engineering system application model for vessels was developed using the ICAM definition method 0 (IDEF0),which is a systematical modeling method.This system model bridges the gap between the data and information flow of every two associated subjects with the major basic research methods and approaches included,which brings the formerly relatively independent subjects together as a whole.The application of this systematic model should facilitate the application of man-machine-environment system engineering in vessels,especially at the conceptual and embodiment design phases.The managers and designers can deal with detailed tasks quickly and efficiently while reducing repetitive work.

  13. Concrete fracture models and applications

    CERN Document Server

    Kumar, Shailendra

    2011-01-01

    Concrete-Fracture Models and Applications provides a basic introduction to nonlinear concrete fracture models. Readers will find a state-of-the-art review on various aspects of the material behavior and development of different concrete fracture models.

  14. Mathematical modeling with multidisciplinary applications

    CERN Document Server

    Yang, Xin-She

    2013-01-01

    Features mathematical modeling techniques and real-world processes with applications in diverse fields Mathematical Modeling with Multidisciplinary Applications details the interdisciplinary nature of mathematical modeling and numerical algorithms. The book combines a variety of applications from diverse fields to illustrate how the methods can be used to model physical processes, design new products, find solutions to challenging problems, and increase competitiveness in international markets. Written by leading scholars and international experts in the field, the

  15. A Habitat-based Wind-Wildlife Collision Model with Application to the Upper Great Plains Region

    Energy Technology Data Exchange (ETDEWEB)

    Forcey, Greg, M.

    2012-08-28

    compared among species, our model outputs provide a convenient and easy landscape-level tool to quickly screen for siting issues at a high level. The model resolution is suitable for state or multi-county siting but users are cautioned against using these models for micrositing. The U.S. Fish and Wildlife Service recently released voluntary land-based wind energy guidelines for assessing impacts of a wind facility to wildlife using a tiered approach. The tiered approach uses an iterative approach for assessing impacts to wildlife in levels of increasing detail from landscape-level screening to site-specific field studies. Our models presented in this paper would be applicable to be used as tools to conduct screening at the tier 1 level and would not be appropriate to complete smaller scale tier 2 and tier 3 level studies. For smaller scale screening ancillary field studies should be conducted at the site-specific level to validate collision predictions.

  16. Investigation of the applicability of a functional programming model to fault-tolerant parallel processing for knowledge-based systems

    Science.gov (United States)

    Harper, Richard

    1989-01-01

    In a fault-tolerant parallel computer, a functional programming model can facilitate distributed checkpointing, error recovery, load balancing, and graceful degradation. Such a model has been implemented on the Draper Fault-Tolerant Parallel Processor (FTPP). When used in conjunction with the FTPP's fault detection and masking capabilities, this implementation results in a graceful degradation of system performance after faults. Three graceful degradation algorithms have been implemented and are presented. A user interface has been implemented which requires minimal cognitive overhead by the application programmer, masking such complexities as the system's redundancy, distributed nature, variable complement of processing resources, load balancing, fault occurrence and recovery. This user interface is described and its use demonstrated. The applicability of the functional programming style to the Activation Framework, a paradigm for intelligent systems, is then briefly described.

  17. From situation modelling to a distributed rule-based platform for situation awareness: an ontological framework for disaster management applications

    OpenAIRE

    Moreira, João Luiz Rebelo

    2015-01-01

    Situation-aware (SA) applications are particularly useful for disaster management. The complex nature of emergency scenarios presents challenges to the development of collaborative and distributed SA solutions. These challenges concern the whole lifecycle, from specification to implementation phases, such as how to model the reaction behavior of a detected situation and how to provide an interoperable situation notification service. In addition, treating unforeseen situations, i.e. situations...

  18. Efficient GIS-based model-driven method for flood risk management and its application in central China

    Science.gov (United States)

    Liu, Y.; Zhou, J.; Song, L.; Zou, Q.; Guo, J.; Wang, Y.

    2014-02-01

    In recent years, an important development in flood management has been the focal shift from flood protection towards flood risk management. This change greatly promoted the progress of flood control research in a multidisciplinary way. Moreover, given the growing complexity and uncertainty in many decision situations of flood risk management, traditional methods, e.g., tight-coupling integration of one or more quantitative models, are not enough to provide decision support for managers. Within this context, this paper presents a beneficial methodological framework to enhance the effectiveness of decision support systems, through the dynamic adaptation of support regarding the needs of the decision-maker. In addition, we illustrate a loose-coupling technical prototype for integrating heterogeneous elements, such as multi-source data, multidisciplinary models, GIS tools and existing systems. The main innovation is the application of model-driven concepts, which put the system in a state of continuous iterative optimization. We define the new system as a model-driven decision support system (MDSS ). Two characteristics that differentiate the MDSS are as follows: (1) it is made accessible to non-technical specialists; and (2) it has a higher level of adaptability and compatibility. Furthermore, the MDSS was employed to manage the flood risk in the Jingjiang flood diversion area, located in central China near the Yangtze River. Compared with traditional solutions, we believe that this model-driven method is efficient, adaptable and flexible, and thus has bright prospects of application for comprehensive flood risk management.

  19. A Unified ASrchitecture Model of Web Applications

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    With the increasing popularity,scale and complexity of web applications,design and development of web applications are becoming more and more difficult,However,the current state of their design and development is characterized by anarchy and ad hoc methodologies,One of the causes of this chaotic situation is that different researchers and designers have different understanding of web applications.In this paper,based on an explicit understanding of web applications,we present a unified architecture model of wed applications,the four-view model,which addresses the analysis and design issues of web applications from four perspectives,namely,logical view,data view,navigation view and presentation view,each addrssing a specific set of concerns of web applications,the purpose of the model is to provide a clear picture of web applications to alleviate the chaotic situation and facilitate its analysis,design and implementation.

  20. Parametric estimation of covariance function in Gaussian-process based Kriging models. Application to uncertainty quantification for computer experiments

    International Nuclear Information System (INIS)

    The parametric estimation of the covariance function of a Gaussian process is studied, in the framework of the Kriging model. Maximum Likelihood and Cross Validation estimators are considered. The correctly specified case, in which the covariance function of the Gaussian process does belong to the parametric set used for estimation, is first studied in an increasing-domain asymptotic framework. The sampling considered is a randomly perturbed multidimensional regular grid. Consistency and asymptotic normality are proved for the two estimators. It is then put into evidence that strong perturbations of the regular grid are always beneficial to Maximum Likelihood estimation. The incorrectly specified case, in which the covariance function of the Gaussian process does not belong to the parametric set used for estimation, is then studied. It is shown that Cross Validation is more robust than Maximum Likelihood in this case. Finally, two applications of the Kriging model with Gaussian processes are carried out on industrial data. For a validation problem of the friction model of the thermal-hydraulic code FLICA 4, where experimental results are available, it is shown that Gaussian process modeling of the FLICA 4 code model error enables to considerably improve its predictions. Finally, for a meta modeling problem of the GERMINAL thermal-mechanical code, the interest of the Kriging model with Gaussian processes, compared to neural network methods, is shown. (author)

  1. Study and Application of Safety Risk Evaluation Model for CO2 Geological Storage Based on Uncertainty Measure Theory

    OpenAIRE

    Hujun He; Yaning Zhao; Xingke Yang; Yaning Gao; Xu Wu

    2015-01-01

    Analyzing showed that the safety risk evaluation for CO2 geological storage had important significance. Aimed at the characteristics of CO2 geological storage safety risk evaluation, drawing on previous research results, rank and order models for safety risk evaluation of CO2 geological storage were put forward based on information entropy and uncertainty measure theory. In this model, the uncertainty problems in safety risk evaluation of CO2 geological storage were solved by qualitative anal...

  2. Estimating Crop Albedo in the Application of a Physical Model Based on the Law of Energy Conservation and Spectral Invariants

    Directory of Open Access Journals (Sweden)

    Jingjing Peng

    2015-11-01

    Full Text Available Albedo characterizes the radiometric interface of land surfaces, especially vegetation, and the atmosphere. Albedo is a critical input to many models, such as crop growth models, hydrological models and climate models. For the extensive attention to crop monitoring, a physical albedo model for crops is developed based on the law of energy conservation and spectral invariants, which is derived from a prior forest albedo model. The model inputs have been efficiently and physically parameterized, including the dependency of albedo on the solar zenith/azimuth angle, the fraction of diffuse skylight in the incident radiance, the canopy structure, the leaf reflectance/transmittance and the soil reflectance characteristics. Both the anisotropy of soil reflectance and the clumping effect of crop leaves at the canopy scale are considered, which contribute to the improvement of the model accuracy. The comparison between the model results and Monte Carlo simulation results indicates that the canopy albedo has high accuracy with an RMSE < 0.005. The validation using ground measurements has also demonstrated the reliability of the model and that it can reflect the interaction mechanism between radiation and the canopy-soil system.

  3. Graph Model Based Indoor Tracking

    DEFF Research Database (Denmark)

    Jensen, Christian Søndergaard; Lu, Hua; Yang, Bin

    2009-01-01

    The tracking of the locations of moving objects in large indoor spaces is important, as it enables a range of applications related to, e.g., security and indoor navigation and guidance. This paper presents a graph model based approach to indoor tracking that offers a uniform data management...... infrastructure for different symbolic positioning technologies, e.g., Bluetooth and RFID. More specifically, the paper proposes a model of indoor space that comprises a base graph and mappings that represent the topology of indoor space at different levels. The resulting model can be used for one or several...... indoor positioning technologies. Focusing on RFID-based positioning, an RFID specific reader deployment graph model is built from the base graph model. This model is then used in several algorithms for constructing and refining trajectories from raw RFID readings. Empirical studies with implementations...

  4. CORBA Based CIMS Application Integration

    Institute of Scientific and Technical Information of China (English)

    1999-01-01

    Common object request broker architecture (CORBA) provides the framework and the mechanism for distributed object operation. It can also be applied to computer integrated manufacturing system (CIMS) application integration. This paper studies the CIMS information service requirement, presents a CORBA based integration approach including the CORBA based CIM information system architecture and the application integration mechanism, and discusses the relationship between CORBA and the CIM application integration platform.

  5. Hyaluronic acid algorithm-based models for assessment of liver ifbrosis:translation from basic science to clinical application

    Institute of Scientific and Technical Information of China (English)

    Zeinab Babaei; Hadi Parsian

    2016-01-01

    BACKGROUND: The estimation of liver ifbrosis is usually dependent on liver biopsy evaluation. Because of its disad-vantages and side effects, researchers try to ifnd non-invasive methods for the assessment of liver injuries. Hyaluronic acid has been proposed as an index for scoring the severity of if-brosis, alone or in algorithm models. The algorithm model in which hyaluronic acid was used as a major constituent was more reliable and accurate in diagnosis than hyaluronic acid alone. This review described various hyaluronic acid algo-rithm-based models for assessing liver ifbrosis. DATA SOURCE: A PubMed database search was performed to identify the articles relevant to hyaluronic acid algorithm-based models for estimating liver ifbrosis. RESULT: The use of hyaluronic acid in an algorithm model is an extra and valuable tool for assessing liver ifbrosis. CONCLUSIONS: Although hyaluronic acid algorithm-based models have good diagnostic power in liver ifbrosis assess-ment, they cannot render the need for liver biopsy obsolete and it is better to use them in parallel with liver biopsy. They can be used when frequent liver biopsy is not possible in situa-tions such as highlighting the efifcacy of treatment protocol for liver ifbrosis.

  6. Simulation-Based Estimation of the Structural Errors-in-Variables Negative Binomial Regression Model with an Application

    OpenAIRE

    Jie Q. Guo; Tong Li

    2001-01-01

    This paper studies the effects and estimation of errors-in-variables negative binomial regression model. We prove that in the presence of measurement errors, in general, maximum likelihood estimator of the overdispersion using the observed data is biased upward. We adopt a structural approach assuming that the distribution of the latent variables is known and propose a simulation-based corrected maximum likelihood estimator and a simulation-based corrected score estimator to estimate the erro...

  7. In vitro permeation models for healthy and compromised skin: The Phospholipid Vesicle-based Permeation Assay (PVPA) for skin applications

    OpenAIRE

    Engesland, André

    2015-01-01

    In vitro models with the ability to estimate drug penetration through healthy and compromised skin may reduce animal testing of drugs and cosmetics to a minimum. The phospholipid vesicle based permeation assay (PVPA) is based on a tight barrier composed of liposomes mimicking cells. It was originally made to mimic the intestinal epithelial barrier and in this project further developed to mimic the stratum corneum barrier of the skin. The lipid composition was changed to better mimic the lipid...

  8. A Capacity Fading Model of Lithium-Ion Battery Cycle Life Based on the Kinetics of Side Reactions for Electric Vehicle Applications

    International Nuclear Information System (INIS)

    Highlights: • Describe the aging mechanism of lithium-ion battery with electrochemical kinetics. • Establish the fading rate equation based on Eyring Equation. • The established equation is applicable to any reaction order. • Integrate the internal kinetics with external degradation characteristics. - Abstract: Battery life prediction is one of the critical issues that restrict the development of electric vehicles. Among the typical battery life models, the mechanism model focusing on the internal physical or electrochemical processes has a stronger theoretical foundation and greater accuracy. The empirical formula, which relies on the simplified mechanism, has a concise model structure and more flexibility in vehicle applications. However, the internal aging mechanism rarely correlates with the external operating characteristics. Based on the summary of the capacity fading mechanism and the reasoning of the internal kinetics of side reactions during the aging process, a lifetime model of the lithium-ion battery is established in this paper. The solutions to the vital parameters based on the external accelerated life testing results are also presented. The testing sample is a manganese oxide lithium-ion battery of 8 Ah. The validation results indicated that the life model established in this paper can describe the capacity fading law of the lithium-ion battery and the operability and accuracy for vehicle applications

  9. Three-parameter-based streamflow elasticity model: application to MOPEX basins in the USA at annual and seasonal scales

    Science.gov (United States)

    Konapala, Goutam; Mishra, Ashok K.

    2016-07-01

    We present a three-parameter streamflow elasticity model as a function of precipitation, potential evaporation, and change in groundwater storage applicable at both seasonal and annual scales. The model was applied to 245 Model Parameter Estimation Experiment (MOPEX) basins spread across the continental USA. The analysis of the modified equation at annual and seasonal scales indicated that the groundwater and surface water storage change contributes significantly to the streamflow elasticity. Overall, in case of annual as well as seasonal water balances, precipitation has higher elasticity values when compared to both potential evapotranspiration and storage changes. The streamflow elasticities show significant nonlinear associations with the climate conditions of the catchments indicating a complex interplay between elasticities and climate variables with substantial seasonal variations.

  10. An Empirical Polarizable Force Field Based on the Classical Drude Oscillator Model: Development History and Recent Applications

    Science.gov (United States)

    2016-01-01

    Molecular mechanics force fields that explicitly account for induced polarization represent the next generation of physical models for molecular dynamics simulations. Several methods exist for modeling induced polarization, and here we review the classical Drude oscillator model, in which electronic degrees of freedom are modeled by charged particles attached to the nuclei of their core atoms by harmonic springs. We describe the latest developments in Drude force field parametrization and application, primarily in the last 15 years. Emphasis is placed on the Drude-2013 polarizable force field for proteins, DNA, lipids, and carbohydrates. We discuss its parametrization protocol, development history, and recent simulations of biologically interesting systems, highlighting specific studies in which induced polarization plays a critical role in reproducing experimental observables and understanding physical behavior. As the Drude oscillator model is computationally tractable and available in a wide range of simulation packages, it is anticipated that use of these more complex physical models will lead to new and important discoveries of the physical forces driving a range of chemical and biological phenomena. PMID:26815602

  11. An Empirical Polarizable Force Field Based on the Classical Drude Oscillator Model: Development History and Recent Applications.

    Science.gov (United States)

    Lemkul, Justin A; Huang, Jing; Roux, Benoît; MacKerell, Alexander D

    2016-05-11

    Molecular mechanics force fields that explicitly account for induced polarization represent the next generation of physical models for molecular dynamics simulations. Several methods exist for modeling induced polarization, and here we review the classical Drude oscillator model, in which electronic degrees of freedom are modeled by charged particles attached to the nuclei of their core atoms by harmonic springs. We describe the latest developments in Drude force field parametrization and application, primarily in the last 15 years. Emphasis is placed on the Drude-2013 polarizable force field for proteins, DNA, lipids, and carbohydrates. We discuss its parametrization protocol, development history, and recent simulations of biologically interesting systems, highlighting specific studies in which induced polarization plays a critical role in reproducing experimental observables and understanding physical behavior. As the Drude oscillator model is computationally tractable and available in a wide range of simulation packages, it is anticipated that use of these more complex physical models will lead to new and important discoveries of the physical forces driving a range of chemical and biological phenomena. PMID:26815602

  12. Generalizability and Applicability of Model-Based Business Process Compliance-Checking Approaches – A State-of-the-Art Analysis and Research Roadmap

    Directory of Open Access Journals (Sweden)

    Jörg Becker

    2012-11-01

    Full Text Available With a steady increase of regulatory requirements for business processes, automation support of compliance management is a field garnering increasing attention in Information Systems research. Several approaches have been developed to support compliance checking of process models. One major challenge for such approaches is their ability to handle different modeling techniques and compliance rules in order to enable widespread adoption and application. Applying a structured literature search strategy, we reflect and discuss compliance-checking approaches in order to provide an insight into their generalizability and evaluation. The results imply that current approaches mainly focus on special modeling techniques and/or a restricted set of types of compliance rules. Most approaches abstain from real-world evaluation which raises the question of their practical applicability. Referring to the search results, we propose a roadmap for further research in model-based business process compliance checking.

  13. Model for water pollution remote sensing based on double scattering and its application in the Zhujiang River outfall

    Institute of Scientific and Technical Information of China (English)

    DENG Ruru; LIU Qinhuo; KE Ruiping; CHENG Lei; LIU Xiaoping

    2004-01-01

    It is a valid route for quantitatively remote sensing on water pollution to build a model according to the physical mechanisms of scattering and absorbing of suspended substance, pollutant, and molecules of water. Remote sensing model for water pollution based on single scattering is simple and easy to be used, but the precision is affected by turbidity of water. The characteristics of the energy composition of multiple scattering, are analyzed and it is proposed that, based on the model of single scattering, ifthe flux of the second scattering is considered additionally, the precision of the modelwill be remarkably improved and the calculation is still very simple. The factor of the second scattering is deduced to build a double scattering model, and the practical arithmetic for the calculation of the model is put forward. The result of applying this model in the water area around the Zhujiang(Pearl) River outfall shows that the precision is obviously improved. The result also shows that the seriously polluted water area is distributed in the northeast of Lingding Sea, the Victoria Bay of Hong Kong, and the Shengzhen Bay.

  14. Modelling Foundations and Applications

    DEFF Research Database (Denmark)

    selected from 81 submissions. Papers on all aspects of MDE were received, including topics such as architectural modelling and product lines, code generation, domain-specic modeling, metamodeling, model analysis and verication, model management, model transformation and simulation. The breadth of topics...

  15. Structural equation modeling methods and applications

    CERN Document Server

    Wang, Jichuan

    2012-01-01

    A reference guide for applications of SEM using Mplus Structural Equation Modeling: Applications Using Mplus is intended as both a teaching resource and a reference guide. Written in non-mathematical terms, this book focuses on the conceptual and practical aspects of Structural Equation Modeling (SEM). Basic concepts and examples of various SEM models are demonstrated along with recently developed advanced methods, such as mixture modeling and model-based power analysis and sample size estimate for SEM. The statistical modeling program, Mplus, is also featured and provides researchers with a

  16. Web-based applications for virtual laboratories

    OpenAIRE

    Bier, H.H.

    2011-01-01

    Web-based applications for academic education facilitate, usually, exchange of multimedia files, while design-oriented domains such as architectural and urban design require additional support in collaborative real-time drafting and modeling. In this context, multi-user interactive interfaces employing game engines as well as Virtual Reality (VR) environments offer a framework within which web-based applications for virtual laboratories have been in the last decade successfully developed and ...

  17. Model-order reduction of magneto-harmonic problems based on POD: application to planar magnetic components

    Science.gov (United States)

    Taylor, Laurent; Henneron, Thomas; Margueron, Xavier; Le Menach, Yvonnick; Le Moigne, Philippe

    2016-04-01

    Predetermination of losses and inductance values in the design phase, is necessary for the development of high-performance magnetic components for power electronics. Numerical modeling, based on the finite element method (FEM) can be used to determine the characteristics of a particular component with a complex geometry in high frequency (HF). These models are very accurate but the computation time required is high compared to analytical models. The model order reduction (MOR) methods can be applied to reduce the computation time while maintaining high accuracy. Nowadays, the proper orthogonal decomposition (POD) is the most popular of MOR approaches. This technique has been applied to study problems in many fields of engineering. In this paper, the POD method is developed to solve magneto-harmonic problems in order to study a HF planar magnetic inductor. Contribution to the topical issue "Numelec 2015-Elected submissions", edited by Adel Razek

  18. MIRAGE: a functional genomics-based approach for metabolic network model reconstruction and its application to cyanobacteria networks.

    Science.gov (United States)

    Vitkin, Edward; Shlomi, Tomer

    2012-01-01

    Genome-scale metabolic network reconstructions are considered a key step in quantifying the genotype-phenotype relationship. We present a novel gap-filling approach, MetabolIc Reconstruction via functionAl GEnomics (MIRAGE), which identifies missing network reactions by integrating metabolic flux analysis and functional genomics data. MIRAGE's performance is demonstrated on the reconstruction of metabolic network models of E. coli and Synechocystis sp. and validated via existing networks for these species. Then, it is applied to reconstruct genome-scale metabolic network models for 36 sequenced cyanobacteria amenable for constraint-based modeling analysis and specifically for metabolic engineering. The reconstructed network models are supplied via standard SBML files. PMID:23194418

  19. A dislocation density based crystal plasticity finite element model: Application to a two-phase polycrystalline HCP/BCC composites

    Science.gov (United States)

    Ardeljan, Milan; Beyerlein, Irene J.; Knezevic, Marko

    2014-05-01

    We present a multiscale model for anisotropic, elasto-plastic, rate- and temperature-sensitive deformation of polycrystalline aggregates to large plastic strains. The model accounts for a dislocation-based hardening law for multiple slip modes and links a single-crystal to a polycrystalline response using a crystal plasticity finite element based homogenization. It is capable of predicting local stress and strain fields based on evolving microstructure including the explicit evolution of dislocation density and crystallographic grain reorientation. We apply the model to simulate monotonic mechanical response of a hexagonal close-packed metal, zirconium (Zr), and a body-centered cubic metal, niobium (Nb), and study the texture evolution and deformation mechanisms in a two-phase Zr/Nb layered composite under severe plastic deformation. The model predicts well the texture in both co-deforming phases to very large plastic strains. In addition, it offers insights into the active slip systems underlying texture evolution, indicating that the observed textures develop by a combination of prismatic, pyramidal, and anomalous basal slip in Zr and primarily {110} slip and secondly {112} slip in Nb.

  20. Model-Based Reasoning

    Science.gov (United States)

    Ifenthaler, Dirk; Seel, Norbert M.

    2013-01-01

    In this paper, there will be a particular focus on mental models and their application to inductive reasoning within the realm of instruction. A basic assumption of this study is the observation that the construction of mental models and related reasoning is a slowly developing capability of cognitive systems that emerges effectively with proper…

  1. Development and application of EEAST: a life cycle based model for use of harvested rainwater and composting toilets in buildings.

    Science.gov (United States)

    Devkota, J; Schlachter, H; Anand, C; Phillips, R; Apul, Defne

    2013-11-30

    Harvested rainwater systems and composting toilets are expected to be an important part of sustainable solutions in buildings. Yet, to this date, a model evaluating their economic and environmental impact has been missing. To address this need, a life cycle based model, EEAST was developed. EEAST was designed to compare the business as usual (BAU) case of using potable water for toilet flushing and irrigation to alternative scenarios of rainwater harvesting and composting toilet based technologies. In EEAST, building characteristics, occupancy, and precipitation are used to size the harvested rainwater and composting toilet systems. Then, life cycle costing and life cycle assessment methods are used to estimate cost, energy, and greenhouse gas (GHG) emission payback periods (PPs) for five alternative scenarios. The scenarios modeled include use of harvested rainwater for toilet flushing, for irrigation, or both; and use of composting toilets with or without harvested rainwater use for irrigation. A sample simulation using EEAST showed that for the office building modeled, the cost PPs were greater than energy PPs which in turn were greater than GHG emission PPs. This was primarily due to energy and emission intensive nature of the centralized water and wastewater infrastructure. The sample simulation also suggested that the composting toilets may have the best performance in all criteria. However, EEAST does not explicitly model solids management and as such may give composting toilets an unfair advantage compared to flush based toilets. EEAST results were found to be very sensitive to cost values used in the model. With the availability of EEAST, life cycle cost, energy, and GHG emissions can now be performed fairly easily by building designers and researchers. Future work is recommended to further improve EEAST and evaluate it for different types of buildings and climates so as to better understand when composting toilets and harvested rainwater systems

  2. Application of model-free kinetics to the study of dehydration of fly ash-based zeolite

    International Nuclear Information System (INIS)

    In the present paper, dehydration kinetics of zeolite Na-A synthesized from fly ash was investigated by means of thermogravimetric analysis. Na-A zeolite was formed from coal fly fash by fusion with sodium hydroxide and succeeding hydrothermal treatment at 100 deg. C after induction period. The model-free kinetic method was applied to calculate the activation energy of the dehydration process of fly ash-based zeolite as a function of conversion and temperature. The Vyazovkin model-free kinetic method also enabled the definition of time, necessary to remove water molecules from the zeolite structure for a given temperature

  3. Fundamental Study on Applicability of Powder-Based 3D Printer for Physical Modeling in Rock Mechanics

    Science.gov (United States)

    Fereshtenejad, Sayedalireza; Song, Jae-Joon

    2016-06-01

    Applications of 3D printing technology become more widespread in many research fields because of its rapid development and valuable capabilities. In rock mechanics and mining engineering, this technology has the potential to become a useful tool that might help implement a number of research studies previously considered impractical. Most commercial 3D printers cannot print prototypes with mechanical properties that match precisely those of natural rock samples. Therefore, some additional enhancements are required for 3D printers to be effectively utilized for rock mechanics applications. In this study, we printed and studied specimens using a powder-based commercial ZPrinter® 450 with ZP® 150 powder and Zb® 63 binder used as raw materials. The specimens printed by this 3D printer exhibited relatively low strength and ductile behavior, implying that it needs further improvements. Hence, we focused on several ways to determine the best combination of printing options and post-processing including the effects of the printing direction, printing layer thickness, binder saturation level, and heating process on the uniaxial compressive strength (UCS) and stress-strain behavior of the printed samples. The suggested procedures have demonstrated their effectiveness by obtaining the printed samples that behave similarly to the natural rocks with low UCS. Although our optimization methods were particularly successful, further improvements are required to expand 3D printer application in the area of rock mechanics.

  4. A Robust Design Applicability Model

    DEFF Research Database (Denmark)

    Ebro, Martin; Lars, Krogstie; Howard, Thomas J.

    2015-01-01

    This paper introduces a model for assessing the applicability of Robust Design (RD) in a project or organisation. The intention of the Robust Design Applicability Model (RDAM) is to provide support for decisions by engineering management considering the relevant level of RD activities. The...

  5. Studying nanotube-based oscillators and their application as memory cells via nanoscale continuum modeling and simulation

    Directory of Open Access Journals (Sweden)

    Shaoping Xiao

    2015-12-01

    Full Text Available A nanoscale continuum model of carbon nanotube-based oscillators is proposed in this paper. In the continuum model, the nanotube is discretized via the meshfree particle method. The atomistic interlayer interaction between the outer and inner tubes is approximated by the interlayer interaction between particles. The mechanical behaviors of oscillators are studied and compared well with molecular dynamics simulation results. The nanotube-based oscillator can be employed to design a nanoelectromechanical system. In this system, two electrodes are attached on the top of the outer tube so that the induced electromagnetic force can overcome the interlayer friction. The mechanisms of such nanoelectromechanical systems as memory cells are also considered.

  6. Land Use Allocation Based on a Multi-Objective Artificial Immune Optimization Model: An Application in Anlu County, China

    OpenAIRE

    Xiaoya Ma; Xiang Zhao

    2015-01-01

    As the main feature of land use planning, land use allocation (LUA) optimization is an important means of creating a balance between the land-use supply and demand in a region and promoting the sustainable utilization of land resources. In essence, LUA optimization is a multi-objective optimization problem under the land use supply and demand constraints in a region. In order to obtain a better sustainable multi-objective LUA optimization solution, the present study proposes a LUA model based...

  7. Methane emissions from floodplains in the Amazon Basin: challenges in developing a process-based model for global applications

    OpenAIRE

    Ringeval, B.; S. Houweling; P. M. van Bodegom; R. Spahni; De Beek, R.; Joos, F.; Röckmann, T.

    2014-01-01

    Tropical wetlands are estimated to represent about 50% of the natural wetland methane (CH4) emissions and explain a large fraction of the observed CH4 variability on timescales ranging from glacial–interglacial cycles to the currently observed year-to-year variability. Despite their importance, however, tropical wetlands are poorly represented in global models aiming to predict global CH4 emissions. This publication documents a first step in the development of a process-base...

  8. Analysis of different model-based approaches for estimating dFRC for real-time application

    OpenAIRE

    Van Drunen, EJ; Chase, JG; Chiew, YS; Shaw, GM; Desaive, Thomas

    2013-01-01

    Background Acute Respiratory Distress Syndrome (ARDS) is characterized by inflammation, filling of the lung with fluid and the collapse of lung units. Mechanical ventilation (MV) is used to treat ARDS using positive end expiratory pressure (PEEP) to recruit and retain lung units, thus increasing pulmonary volume and dynamic functional residual capacity (dFRC) at the end of expiration. However, simple, non-invasive methods to estimate dFRC do not exist. Methods Four model-based methods for est...

  9. Land Use Allocation Based on a Multi-Objective Artificial Immune Optimization Model: An Application in Anlu County, China

    Directory of Open Access Journals (Sweden)

    Xiaoya Ma

    2015-11-01

    Full Text Available As the main feature of land use planning, land use allocation (LUA optimization is an important means of creating a balance between the land-use supply and demand in a region and promoting the sustainable utilization of land resources. In essence, LUA optimization is a multi-objective optimization problem under the land use supply and demand constraints in a region. In order to obtain a better sustainable multi-objective LUA optimization solution, the present study proposes a LUA model based on the multi-objective artificial immune optimization algorithm (MOAIM-LUA model. The main achievements of the present study are as follows: (a the land-use supply and demand factors are analyzed and the constraint conditions of LUA optimization problems are constructed based on the analysis framework of the balance between the land use supply and demand; (b the optimization objectives of LUA optimization problems are defined and modeled using ecosystem service value theory and land rent and price theory; and (c a multi-objective optimization algorithm is designed for solving multi-objective LUA optimization problems based on the novel immune clonal algorithm (NICA. On the basis of the aforementioned achievements, MOAIM-LUA was applied to a real case study of land-use planning in Anlu County, China. Compared to the current land use situation in Anlu County, optimized LUA solutions offer improvements in the social and ecological objective areas. Compared to the existing models, such as the non-dominated sorting genetic algorithm-II, experimental results demonstrate that the model designed in the present study can obtain better non-dominated solution sets and is superior in terms of algorithm stability.

  10. Model-based real-time control for laser induced thermal therapy with applications to prostate cancer treatment

    Science.gov (United States)

    Feng, Yusheng; Fuentes, David; Stafford, R. Jason; Oden, J. Tinsley

    2009-02-01

    In this paper, we present a model-based predictive control system that is capable of capturing physical and biological variations of laser-tissue interaction as well as heterogeneity in real-time during laser induced thermal therapy (LITT). Using a three-dimensional predictive bioheat transfer model, which is built based on regular magnetic resonance imaging (MRI) anatomic scan and driven by imaging data produced by real-time magnetic resonance temperature imaging (MRTI), the computational system provides a regirous real-time predictive control during surgical operation process. The unique feature of the this system is its ability for predictive control based on validated model with high precision in real-time, which is made possible by implementation of efficient parallel algorithms. The major components of the current computational systems involves real-time finite element solution of the bioheat transfer induced by laser-tissue interaction, solution module of real-time calibration problem, optimal laser source control, goal-oriented error estimation applied to the bioheat transfer equation, and state-of-the-art imaging process module to characterize the heterogeneous biological domain. The system was tested in vivo in a canine animal model in which an interstitial laser probe was placed in the prostate region and the desired treatment outcome in terms of ablation temperature and damage zone were achieved. Using the guidance of the predictive model driven by real-time MRTI data while applying the optimized laser heat source has the potential to provide unprecedented control over the treatment outcome for laser ablation.

  11. Different-source gas emission prediction model of working face based on BP artificial neural network and its application

    Energy Technology Data Exchange (ETDEWEB)

    Zhu, H.; Chang, W.; Zhang, B. [China University of Mining and Technology, Beijing (China)

    2007-05-15

    Back-propagation (BP) neural network analysis based on the difference- source gas emission quantity prediction theory was applied to predict the quantity of gas emitted from the coal seam being mined, the neighbouring coal seam and the goaf of the working face. Three separate gas emission prediction neural network models were established for these. The prediction model of the coal seam being mined was made up of three layers and nine parameters; that of the neighbouring coal seam was made up of three layers and eight parameters; and that of the goaf of three layers and four parameters. The difference-source gas emission prediction model can greatly improve prediction accuracy. BP neural network analysis using Matlab software was applied in a coal mine. 10 refs., 2 figs., 3 tabs.

  12. Diagnosis of dynamic systems based on explicit and implicit behavioural models: an application to gas turbines in Esprit Project Tiger

    Energy Technology Data Exchange (ETDEWEB)

    Trave-Massuyes, L. [Centre National de la Recherche Scientifique (CNRS), 31 - Toulouse (France); Milne, R.

    1995-12-31

    We are interested in the monitoring and diagnosis of dynamic systems. In our work, we are combining explicit temporal models of the behaviour of a dynamic system with implicit behavioural models supporting model based approaches. This work is drive by the needs of and applied to, two gas turbines of very different size and power. In this paper we describe the problems of building systems for these domains and illustrate how we have developed a system where these two approaches complement each other to provide a comprehensive fault detection and diagnosis system. We also explore the strengths and weaknesses of each approach. The work described here is currently working continuously, on line to a gas turbine in a major chemical plant. (author) 24 refs.

  13. Physiologically based pharmacokinetic models of small molecules and therapeutic antibodies: a mini-review on fundamental concepts and applications.

    Science.gov (United States)

    Ferl, Gregory Z; Theil, Frank-Peter; Wong, Harvey

    2016-03-01

    The mechanisms of absorption, distribution, metabolism and elimination of small and large molecule therapeutics differ significantly from one another and can be explored within the framework of a physiologically based pharmacokinetic (PBPK) model. This paper briefly reviews fundamental approaches to PBPK modeling, in which drug kinetics within tissues and organs are explicitly represented using physiologically meaningful parameters. The differences in PBPK models applied to small/large molecule drugs are highlighted, thus elucidating differences in absorption, distribution and elimination properties between these two classes of drugs in a systematic manner. The absorption of small and large molecules differs with respect to their common extravascular routes of delivery (oral versus subcutaneous). The role of the lymphatic system in drug distribution, and the involvement of tissues as sites of elimination (through catabolism and target mediated drug disposition) are unique features of antibody distribution and elimination that differ from small molecules, which are commonly distributed into the tissues but are eliminated primarily by liver metabolism. Fundamental differences exist in the ability to predict human pharmacokinetics based upon preclinical data due to differing mechanisms governing small and large molecule disposition. These differences have influence on the evolving utilization of PBPK modeling in the discovery and development of small and large molecule therapeutics. Copyright © 2015 John Wiley & Sons, Ltd. PMID:26461173

  14. Model-Based Pseudo-Quad-Pol Reconstruction from Compact Polarimetry and Its Application to Oil-Spill Observation

    Directory of Open Access Journals (Sweden)

    Junjun Yin

    2015-01-01

    Full Text Available Compact polarimetry is an effective imaging mode for wide area observation, especially for the open ocean. In this study, we propose a new method for pseudo-quad-polarization reconstruction from compact polarimetry based on the three-component decomposition. By using the decomposed powers, the reconstruction model is established as a power-weighted model. Further, the phase of the copolarized correlation is taken into consideration. The phase of double-bounce scattering is closer to π than to 0, while the phase of surface scattering is closer to 0 than to π. By considering the negative (double-bounce reflection and positive (surface reflection copolarized correlation, the reconstruction model for full polarimetry has a good consistency with the real polarimetric SAR data. L-band ALOS/PALSAR-1 fully polarimetric data acquired on August 27, 2006, over an oil-spill area are used for demonstration. Reconstruction performance is evaluated with a set of typical polarimetric oil-spill indicators. Quantitative comparison is given. Results show that the proposed model-based method is of great potential for oil-spill observation.

  15. GIS-Based (W+-W-) Weight of Evidence Model and Its Application to Gold Resources Assessment in Abitibi, Canada

    Institute of Scientific and Technical Information of China (English)

    2006-01-01

    The weight of evidence (WofE) model has been widely used for mineral potential mapping.During the conversion of a multiclass map into a binary map a lot of mineralization information is artificially added or lost because the generalization of the class within the cumulative distance interval to a linear feature is based on a maximum contrast, which matches a cumulative distance interval. Additionally,some categorical data evidence cannot be generated by this method because a maximum contrast does not exist. In this article, an alternative (W+ -W- )-based WofE model is proposed. In this model, the "(W+ -W- ) greater than zero or not" is used as a criterion to reclassify the corresponding categorical class into a presence or absence class to convert a multiclass map into a binary map. This model can be applied to both categorical data and successive data. The latter can be operated as categorical data. The W+ and W- of the generated binary maps can be recalculated, and several binary maps can be integrated on the condition that the reclassified binary evidences are conditionally independent of each other. This method effectively reduces artificial data and both nominal and categorical data can be operated. A case study of gold potential mapping in the Abitibi area, Ontario, Canada, shows that the gold potential map by the (W+ -W- ) model displays a smaller potential area but a higher posterior probability (POP),whereas the potential map by the traditional (W+ -W- ) model exhibits a larger potential area but a lower POP.

  16. A Statistically-Based Low-Level Cloud Scheme and Its Tentative Application in a General Circulation Model

    Institute of Scientific and Technical Information of China (English)

    DAI Fushan; YU Rucong; ZHANG Xuehong; YU Yongqiang

    2005-01-01

    In this study, a statistical cloud scheme is first introduced and coupled with a first-order turbulence scheme with second-order turbulence moments parameterized by the timescale of the turbulence dissipation and the vertical turbulent diffusion coefficient. Then the ability of the scheme to simulate cloud fraction at different relative humidity, vertical temperature profile, and the timescale of the turbulent dissipation is examined by numerical simulation. It is found that the simulated cloud fraction is sensitive to the parameter used in the statistical cloud scheme and the timescale of the turbulent dissipation. Based on the analyses, the introduced statistical cloud scheme is modified. By combining the modified statistical cloud scheme with a boundary layer cumulus scheme, a new statistically-based low-level cloud scheme is proposed and tentatively applied in NCAR (National Center for Atmospheric Research) CCM3 (Community Climate Model version3). It is found that the simulation of low-level cloud fraction is markedly improved and the centers with maximum low-level cloud fractions are well simulated in the cold oceans off the western coasts with the statistically-based low-level cloud scheme applied in CCM3. It suggests that the new statistically-based low-level cloud scheme has a great potential in the general circulation model for improving the low-level cloud parameterization.

  17. Comprehensive distributed-parameters modeling and experimental validation of microcantilever-based biosensors with an application to ultrasmall biological species detection

    International Nuclear Information System (INIS)

    Nanotechnological advancements have made a great contribution in developing label-free and highly sensitive biosensors. The detection of ultrasmall adsorbed masses has been enabled by such sensors which transduce molecular interaction into detectable physical quantities. More specifically, microcantilever-based biosensors have caught widespread attention for offering a label-free, highly sensitive and inexpensive platform for biodetection. Although there are a lot of studies investigating microcantilever-based sensors and their biological applications, a comprehensive mathematical modeling and experimental validation of such devices providing a closed form mathematical framework is still lacking. In almost all of the studies, a simple lumped-parameters model has been proposed. However, in order to have a precise biomechanical sensor, a comprehensive model is required being capable of describing all phenomena and dynamics of the biosensor. Therefore, in this study, an extensive distributed-parameters modeling framework is proposed for the piezoelectric microcantilever-based biosensor using different methodologies for the purpose of detecting an ultrasmall adsorbed mass over the microcantilever surface. An optimum modeling methodology is concluded and verified with the experiment. This study includes three main parts. In the first part, the Euler–Bernoulli beam theory is used to model the nonuniform piezoelectric microcantilever. Simulation results are obtained and presented. The same system is then modeled as a nonuniform rectangular plate. The simulation results are presented describing model's capability in the detection of an ultrasmall mass. Finally the last part presents the experimental validation verifying the modeling results. It was shown that plate modeling predicts the real situation with a degree of precision of 99.57% whereas modeling the system as an Euler–Bernoulli beam provides a 94.45% degree of precision. The detection of ultrasmall

  18. Marginal regression models for clustered count data based on zero-inflated Conway-Maxwell-Poisson distribution with applications.

    Science.gov (United States)

    Choo-Wosoba, Hyoyoung; Levy, Steven M; Datta, Somnath

    2016-06-01

    Community water fluoridation is an important public health measure to prevent dental caries, but it continues to be somewhat controversial. The Iowa Fluoride Study (IFS) is a longitudinal study on a cohort of Iowa children that began in 1991. The main purposes of this study (http://www.dentistry.uiowa.edu/preventive-fluoride-study) were to quantify fluoride exposures from both dietary and nondietary sources and to associate longitudinal fluoride exposures with dental fluorosis (spots on teeth) and dental caries (cavities). We analyze a subset of the IFS data by a marginal regression model with a zero-inflated version of the Conway-Maxwell-Poisson distribution for count data exhibiting excessive zeros and a wide range of dispersion patterns. In general, we introduce two estimation methods for fitting a ZICMP marginal regression model. Finite sample behaviors of the estimators and the resulting confidence intervals are studied using extensive simulation studies. We apply our methodologies to the dental caries data. Our novel modeling incorporating zero inflation, clustering, and overdispersion sheds some new light on the effect of community water fluoridation and other factors. We also include a second application of our methodology to a genomic (next-generation sequencing) dataset that exhibits underdispersion. PMID:26575079

  19. Fusion strategies for selecting multiple tuning parameters for multivariate calibration and other penalty based processes: A model updating application for pharmaceutical analysis.

    Science.gov (United States)

    Tencate, Alister J; Kalivas, John H; White, Alexander J

    2016-05-19

    New multivariate calibration methods and other processes are being developed that require selection of multiple tuning parameter (penalty) values to form the final model. With one or more tuning parameters, using only one measure of model quality to select final tuning parameter values is not sufficient. Optimization of several model quality measures is challenging. Thus, three fusion ranking methods are investigated for simultaneous assessment of multiple measures of model quality for selecting tuning parameter values. One is a supervised learning fusion rule named sum of ranking differences (SRD). The other two are non-supervised learning processes based on the sum and median operations. The effect of the number of models evaluated on the three fusion rules are also evaluated using three procedures. One procedure uses all models from all possible combinations of the tuning parameters. To reduce the number of models evaluated, an iterative process (only applicable to SRD) is applied and thresholding a model quality measure before applying the fusion rules is also used. A near infrared pharmaceutical data set requiring model updating is used to evaluate the three fusion rules. In this case, calibration of the primary conditions is for the active pharmaceutical ingredient (API) of tablets produced in a laboratory. The secondary conditions for calibration updating is for tablets produced in the full batch setting. Two model updating processes requiring selection of two unique tuning parameter values are studied. One is based on Tikhonov regularization (TR) and the other is a variation of partial least squares (PLS). The three fusion methods are shown to provide equivalent and acceptable results allowing automatic selection of the tuning parameter values. Best tuning parameter values are selected when model quality measures used with the fusion rules are for the small secondary sample set used to form the updated models. In this model updating situation, evaluation of

  20. A model for assessment of telemedicine applications

    DEFF Research Database (Denmark)

    Kidholm, Kristian; Ekeland, Anne Granstrøm; Jensen, Lise Kvistgaard;

    2012-01-01

    Telemedicine applications could potentially solve many of the challenges faced by the healthcare sectors in Europe. However, a framework for assessment of these technologies is need by decision makers to assist them in choosing the most efficient and cost-effective technologies. Therefore in 2009...... the European Commission initiated the development of a framework for assessing telemedicine applications, based on the users' need for information for decision making. This article presents the Model for ASsessment of Telemedicine applications (MAST) developed in this study....

  1. Multilevel Models Applications Using SAS

    CERN Document Server

    Wang, Jichuan; Fisher, James

    2011-01-01

    This book covers a broad range of topics about multilevel modeling. The goal is to help readersto understand the basic concepts, theoretical frameworks, and application methods of multilevel modeling. Itis at a level also accessible to non-mathematicians, focusing on the methods and applications of various multilevel models and using the widely used statistical software SAS®.Examples are drawn from analysis of real-world research data.

  2. Ontology-based application integration

    CERN Document Server

    Paulheim, Heiko

    2011-01-01

    Ontology-based Application Integration introduces UI-level (User Interface Level) application integration and discusses current problems which can be remedied by using ontologies. It shows a novel approach for applying ontologies in system integration. While ontologies have been used for integration of IT systems on the database and on the business logic layer, integration on the user interface layer is a novel field of research. This book also discusses how end users, not only developers, can benefit from semantic technologies. Ontology-based Application Integration presents the development o

  3. The development and application of a risk-based prioritization model for the Oak Ridge Environmental Restoration Program

    International Nuclear Information System (INIS)

    The Oak Ridge Environmental Restoration (ER) Program developed and implemented the Environmental Restoration Benefit Assessment Matrix (ERBAM) early in 1994 to provide a simple, efficient process for prioritizing and justifying fiscal budget decisions for a diverse set of activities. The decision to develop a methodology for prioritizing sites was necessitated by the large number of buildings and areas managed by the DOE Oak Ridge Field Office and the finite resources available to address these areas. The ERBAM was based on the Integrated Resource Management System prioritization methodology historically used by the United States Department of Energy (DOE) and Lockheed Martin Energy Systems, Inc., to rank compliance and operational activities. To develop the matrix, ER Program management, working with federal and state regulators, agreed on impact criteria that balance the major objectives within the ER Program: protection of public health, protection of the environment, protection of on-site workers, consideration of stakeholder/community preference, achievement of ER mission, and optimization of cost efficiency. Lessons learned from the initial application of the matrix were used to make refinements and improvements in the methodology. A standard set of assumptions (both overall and categoric) and a prioritization board, consisting of top level DOE and Lockheed Martin Energy Systems, Inc., managers along with federal and state regulatory representatives, were established to facilitate consistent application. Current and future improvements include a method to incorporate existing quantitative risk data and facilitate increased efficiency in applying baseline cost data and approved funding levels to the prioritized output. Application of the prioritization methodology yields a prioritized list of all work activities within the programs' work breakdown structure

  4. Development of a pyrolysis waste recovery model with designs, test plans, and applications for space-based habitats

    Science.gov (United States)

    Roberson, Bobby J.

    1992-01-01

    Extensive literature searches revealed the numerous advantages of using pyrolysis as a means of recovering usable resources from inedible plant biomass, paper, plastics, other polymers, and human waste. A possible design of a pyrolysis reactor with test plans and applications for use on a space-based habitat are proposed. The proposed system will accommodate the wastes generated by a four-person crew while requiring solar energy as the only power source. Waste materials will be collected and stored during the 15-day lunar darkness periods. Resource recovery will occur during the daylight periods. Usable gases such as methane and hydrogen and a solid char will be produced while reducing the mass and volume of the waste to almost infinitely small levels. The system will be operated economically, safely, and in a non-polluting manner.

  5. An inverse vibration-based approach towards modelling and damage identification in nonlinearly vibrating structures. Application for delamination detection in a composite beam

    International Nuclear Information System (INIS)

    This study explores the possibilities for inverse analysis and modelling from data of a nonlinearly vibrating structure. We are suggesting a statistical approach based on singular spectrum analysis (SSA). The method is based on a free decay response, when the structure is given an initial disturbance and is left to vibrate on its own. The measured vibration response is decomposed into new variables, the principal components, which are used to uncover oscillatory patterns in the structural response. In this study an application of the methodology for the purposes of delamination detection in a composite beam is explored.

  6. Distributed Parameter Modelling Applications

    DEFF Research Database (Denmark)

    Sales-Cruz, Mauricio; Cameron, Ian; Gani, Rafiqul

    2011-01-01

    development of a short-path evaporator. The oil shale processing problem illustrates the interplay amongst particle flows in rotating drums, heat and mass transfer between solid and gas phases. The industrial application considers the dynamics of an Alberta-Taciuk processor, commonly used in shale oil and oil...... the steady state, distributed behaviour of a short-path evaporator....

  7. State-based modeling of continuous human-integrated systems: An application to air traffic separation assurance

    International Nuclear Information System (INIS)

    A method for modeling the safety of human-integrated systems that have continuous dynamics is introduced. The method is intended to supplement more detailed reliability-based methods. Assumptions for the model are defined such that the model is demonstrably complete, enabling it to yield a set of key agent characteristics. These key characteristics identify a sufficient set of characteristics that can be used to establish the safety of particular system configurations. The method is applied for the analysis of the safety of strategic and tactical separation assurance algorithms for the next generation air transportation system. It is shown that the key characteristics for this problem include the ability of agents (human or automated) to identify configurations that can enable intense transitions from a safe to unsafe state. However, the most technologically advanced algorithm for separation assurance does not currently attempt to identify such configurations. It is also discussed how, although the model is in a form that lends itself to quantitative evaluations, such evaluations are complicated by the difficulty of accurately quantifying human error probabilities.

  8. Establishment and application of drilling sealing model in the spherical grouting mode based on the loosing-circle theory

    Institute of Scientific and Technical Information of China (English)

    Hao Zhiyong; Lin Baiquan; Gao Yabin; Cheng Yanying

    2012-01-01

    There are quite a few studies that have been done on borehole sealing theory both domestically and internationally.The existing researches usually consider drilling of the surroundings as a dense homogeneous elastic body which does not meet the characteristics of real drilling of the fractured body.Based on the loosing-circle theory and analyses of the surrounding rock stress field,cracks and seepage fields,combined with Newtonian fluid spherical grouting model,we deduced the dynamic relationship between the seepage coefficient and rock or grouting parameters of the drilling sealing fluid mode of spherical fissure grouting.In this experiment,mucus was injected in the simulated coal seam and the permeability coefficient of the sealing body was calculated by using the model.To verify the validity of the model,the calculated sealing body number was compared with the extreme negative pressure that the sealing body could withstand.The theoretical model revealed the drilling sealing fluid mechanism,provided a method for the quantitative calculation of the drilling sealing fluid effect by grouting mode and a reference for the subsequent research of sealing mechanism.

  9. State-based modeling of continuous human-integrated systems: An application to air traffic separation assurance

    Energy Technology Data Exchange (ETDEWEB)

    Landry, Steven J., E-mail: slandry@purdue.ed [School of Industrial Engineering, Purdue University, 315 N. Grant St., West Lafayette, IN 47907 (United States); Lagu, Amit; Kinnari, Jouko [School of Industrial Engineering, Purdue University, 315 N. Grant St., West Lafayette, IN 47907 (United States)

    2010-04-15

    A method for modeling the safety of human-integrated systems that have continuous dynamics is introduced. The method is intended to supplement more detailed reliability-based methods. Assumptions for the model are defined such that the model is demonstrably complete, enabling it to yield a set of key agent characteristics. These key characteristics identify a sufficient set of characteristics that can be used to establish the safety of particular system configurations. The method is applied for the analysis of the safety of strategic and tactical separation assurance algorithms for the next generation air transportation system. It is shown that the key characteristics for this problem include the ability of agents (human or automated) to identify configurations that can enable intense transitions from a safe to unsafe state. However, the most technologically advanced algorithm for separation assurance does not currently attempt to identify such configurations. It is also discussed how, although the model is in a form that lends itself to quantitative evaluations, such evaluations are complicated by the difficulty of accurately quantifying human error probabilities.

  10. Modeling of surface roughness scattering in nanowires based on atomistic wave function: Application to hole mobility in rectangular germanium nanowires

    Science.gov (United States)

    Tanaka, Hajime; Suda, Jun; Kimoto, Tsunenobu

    2016-04-01

    The authors present a calculation model of surface roughness scattering (SRS) in nanowires (NWs) based on atomistic description of electronic states by an s p3d5s* tight-binding scheme, and then this model is applied to hole transport in rectangular cross-sectional germanium (Ge) NWs. In this SRS model, the change of electronic band structures due to width or height reduction is first computed, and then it is expressed using an equivalent potential near the surface. The perturbation corresponding to a surface roughness is calculated from this equivalent potential. Using the aforementioned SRS model, hole mobility in Ge NWs was computed taking into account phonon scattering and SRS. The impacts of SRS on hole mobility in Ge NWs were analyzed, focusing on the valence band structure and hole states of NWs. The main results are as follows. At low hole density, the impacts of SRS are strongly dependent on NW geometry, and Ge NWs with high phonon-limited hole mobility, such as rectangular cross-sectional [110]-oriented NWs with large height along the [001] direction and square cross-sectional [111]-oriented NWs, tend to be less affected by SRS. At high hole density, however, the geometry dependence of hole mobility becomes weaker. These are understood from the nature of hole states and the valence band structure.

  11. Application of a Theory and Simulation based Convective Boundary Mixing model for AGB Star Evolution and Nucleosynthesis

    CERN Document Server

    Battino, U; Ritter, C; Herwig, F; Denisenkov, P; Hartogh, J W Den; Trappitsch, R; Hirschi, R; Freytag, B; Thielemann, F; Paxton, B

    2016-01-01

    The s-process nucleosynthesis in Asymptotic Giant Branch (AGB) stars depends on the modeling of convective boundaries. We present models and s-process simulations that adopt a treatment of convective boundaries based on the results of hydrodynamic simulations and on the theory of mixing due to gravity waves in the vicinity of convective boundaries. Hydrodynamics simulations suggest the presence of convective boundary mixing (CBM) at the bottom of the thermal pulse-driven convective zone. Similarly, convection-induced mixing processes are proposed for the mixing below the convective envelope during third dredge-up where the 13C pocket for the s process in AGB stars forms. In this work we apply a CBM model motivated by simulations and theory to models with initial mass M = 2 and M = 3M?, and with initial metal content Z = 0:01 and Z = 0:02. As reported previously, the He-intershell abundance of 12C and 16O are increased by CBM at the bottom of pulse-driven convection zone. This mixing is affecting the 22Ne(alph...

  12. Models for Dynamic Applications

    DEFF Research Database (Denmark)

    Sales-Cruz, Mauricio; Morales Rodriguez, Ricardo; Heitzig, Martina;

    2011-01-01

    This chapter covers aspects of the dynamic modelling and simulation of several complex operations that include a controlled blending tank, a direct methanol fuel cell that incorporates a multiscale model, a fluidised bed reactor, a standard chemical reactor and finally a polymerisation reactor. T...

  13. Modeling of a Membrane Based Humidifier for Fuel Cell Applications Subject to End-Of-Life Conditions

    DEFF Research Database (Denmark)

    Nielsen, Mads Pagh; Olesen, Anders Christian; Menard, Alan

    2014-01-01

    Proton Exchange Membrane (PEM) Fuel Cell Stacks efficiently convert the chemical energy in hydrogen to electricity through electrochemical reactions occurring on either side of a proton conducting electrolyte. This is a promising and very robust energy conversion process which can be used in many......-based water permeable membrane. Results are presented at nominal BOL-conditions and extreme EOL-conditions. A detailed sub-model incorporating the water absorption/desorption kinetics of Nafion and a novel and accurate representation of the diffusion coefficient of water in Nafion was implemented. The...

  14. Modeling and preliminary characterization of passive, wireless temperature sensors for harsh environment applications based on periodic structures

    Science.gov (United States)

    Delfin Manriquez, Diego I.

    Wireless temperature sensing has attained significant attention in recent years due to the increasing need to develop reliable and affordable sensing solutions for energy conversion systems and other harsh environment applications. The development of next generation sensors for energy production processing parameters, such as temperature and pressure, can result in better performance of the system. Particularly, continuous temperature monitoring in energy conversion systems can result in enhancements such as better system integrity, less pollution and higher thermal efficiencies. However, the conditions experienced in these system components hinder the performance of current solutions due to the presence of semi-conductor materials and welded joints. Additionally, the use of wired systems can result in complex wiring networks, increasing the cost of installation, maintenance and sensor replacement. Therefore, next generation sensing solutions must be developed to overcome current challenges in systems where adverse conditions are present. This research project proposes two novel passive, wireless temperature sensor designs based on concepts of guided mode resonance filters (GMRF) and metamaterials. For the GMRF, a tri-layer structure using a metallic encasing and a circular aperture grating layer was developed to have a resonance frequency of 10 GHz. While for the metamaterial-based sensor a continuation of previous work was presented by utilizing a dielectric substrate and an array of commercially available metallic washers divided in two layers. For both designs, High Frequency Structure Simulator (HFSS) from ANSYSRTM was employed to assess the feasibility of the sensor as well as to optimize the geometry and guide the fabrication process. A systematic approach consisting of evaluating the unit cell, then assessing the number of periods needed, and finally characterizing the response of the final sensor was followed for each case. After the modeling process was

  15. Application of a Theory and Simulation-based Convective Boundary Mixing Model for AGB Star Evolution and Nucleosynthesis

    Science.gov (United States)

    Battino, U.; Pignatari, M.; Ritter, C.; Herwig, F.; Denisenkov, P.; Den Hartogh, J. W.; Trappitsch, R.; Hirschi, R.; Freytag, B.; Thielemann, F.; Paxton, B.

    2016-08-01

    The s-process nucleosynthesis in Asymptotic giant branch (AGB) stars depends on the modeling of convective boundaries. We present models and s-process simulations that adopt a treatment of convective boundaries based on the results of hydrodynamic simulations and on the theory of mixing due to gravity waves in the vicinity of convective boundaries. Hydrodynamics simulations suggest the presence of convective boundary mixing (CBM) at the bottom of the thermal pulse-driven convective zone. Similarly, convection-induced mixing processes are proposed for the mixing below the convective envelope during third dredge-up (TDU), where the {}13{{C}} pocket for the s process in AGB stars forms. In this work, we apply a CBM model motivated by simulations and theory to models with initial mass M = 2 and M=3 {M}ȯ , and with initial metal content Z = 0.01 and Z = 0.02. As reported previously, the He-intershell abundances of {}12{{C}} and {}16{{O}} are increased by CBM at the bottom of the pulse-driven convection zone. This mixing is affecting the {}22{Ne}(α, n){}25{Mg} activation and the s-process efficiency in the {}13{{C}}-pocket. In our model, CBM at the bottom of the convective envelope during the TDU represents gravity wave mixing. Furthermore, we take into account the fact that hydrodynamic simulations indicate a declining mixing efficiency that is already about a pressure scale height from the convective boundaries, compared to mixing-length theory. We obtain the formation of the {}13{{C}}-pocket with a mass of ≈ {10}-4 {M}ȯ . The final s-process abundances are characterized by 0.36\\lt [{{s}}/{Fe}]\\lt 0.78 and the heavy-to-light s-process ratio is -0.23\\lt [{hs}/{ls}]\\lt 0.45. Finally, we compare our results with stellar observations, presolar grain measurements and previous work.

  16. A Space-Time Network-Based Modeling Framework for Dynamic Unmanned Aerial Vehicle Routing in Traffic Incident Monitoring Applications

    Science.gov (United States)

    Zhang, Jisheng; Jia, Limin; Niu, Shuyun; Zhang, Fan; Tong, Lu; Zhou, Xuesong

    2015-01-01

    It is essential for transportation management centers to equip and manage a network of fixed and mobile sensors in order to quickly detect traffic incidents and further monitor the related impact areas, especially for high-impact accidents with dramatic traffic congestion propagation. As emerging small Unmanned Aerial Vehicles (UAVs) start to have a more flexible regulation environment, it is critically important to fully explore the potential for of using UAVs for monitoring recurring and non-recurring traffic conditions and special events on transportation networks. This paper presents a space-time network- based modeling framework for integrated fixed and mobile sensor networks, in order to provide a rapid and systematic road traffic monitoring mechanism. By constructing a discretized space-time network to characterize not only the speed for UAVs but also the time-sensitive impact areas of traffic congestion, we formulate the problem as a linear integer programming model to minimize the detection delay cost and operational cost, subject to feasible flying route constraints. A Lagrangian relaxation solution framework is developed to decompose the original complex problem into a series of computationally efficient time-dependent and least cost path finding sub-problems. Several examples are used to demonstrate the results of proposed models in UAVs’ route planning for small and medium-scale networks. PMID:26076404

  17. Collective estimation of multiple bivariate density functions with application to angular-sampling-based protein loop modeling

    KAUST Repository

    Maadooliat, Mehdi

    2015-10-21

    This paper develops a method for simultaneous estimation of density functions for a collection of populations of protein backbone angle pairs using a data-driven, shared basis that is constructed by bivariate spline functions defined on a triangulation of the bivariate domain. The circular nature of angular data is taken into account by imposing appropriate smoothness constraints across boundaries of the triangles. Maximum penalized likelihood is used to fit the model and an alternating blockwise Newton-type algorithm is developed for computation. A simulation study shows that the collective estimation approach is statistically more efficient than estimating the densities individually. The proposed method was used to estimate neighbor-dependent distributions of protein backbone dihedral angles (i.e., Ramachandran distributions). The estimated distributions were applied to protein loop modeling, one of the most challenging open problems in protein structure prediction, by feeding them into an angular-sampling-based loop structure prediction framework. Our estimated distributions compared favorably to the Ramachandran distributions estimated by fitting a hierarchical Dirichlet process model; and in particular, our distributions showed significant improvements on the hard cases where existing methods do not work well.

  18. A microcomputer-based model of radionuclide spills and discharge plumes for application to the Great Lakes

    International Nuclear Information System (INIS)

    This report describes the implementation of a spill model for a shore-based contamination source on Lake Ontario, and its extensions (implemented and proposed) to situations other than that for which it was originally intended. The original programs were developed to model the transport of radionuclide contaminants in the wind-driven shore currents of the Lake. The information relevant to the simulation consists of depth contours adjacent to the shore, and wind history at half-day intervals from the previous thirty days. The codes have been written in the programming language C, for IBM PC or PC compatibles, preferably with the arithmetic co-processor. Much of the code has been written using the Halo graphics package. Features of the simulation codes include: reasonable processing speed and medium -resolution visual display, and variations on the mode of wind history input graphical output display. The model is suitable in its current form only for straight or concave shorelines. Extensions to convex contours and to reactive pollutants are under development

  19. A Space-Time Network-Based Modeling Framework for Dynamic Unmanned Aerial Vehicle Routing in Traffic Incident Monitoring Applications

    Directory of Open Access Journals (Sweden)

    Jisheng Zhang

    2015-06-01

    Full Text Available It is essential for transportation management centers to equip and manage a network of fixed and mobile sensors in order to quickly detect traffic incidents and further monitor the related impact areas, especially for high-impact accidents with dramatic traffic congestion propagation. As emerging small Unmanned Aerial Vehicles (UAVs start to have a more flexible regulation environment, it is critically important to fully explore the potential for of using UAVs for monitoring recurring and non-recurring traffic conditions and special events on transportation networks. This paper presents a space-time network- based modeling framework for integrated fixed and mobile sensor networks, in order to provide a rapid and systematic road traffic monitoring mechanism. By constructing a discretized space-time network to characterize not only the speed for UAVs but also the time-sensitive impact areas of traffic congestion, we formulate the problem as a linear integer programming model to minimize the detection delay cost and operational cost, subject to feasible flying route constraints. A Lagrangian relaxation solution framework is developed to decompose the original complex problem into a series of computationally efficient time-dependent and least cost path finding sub-problems. Several examples are used to demonstrate the results of proposed models in UAVs’ route planning for small and medium-scale networks.

  20. A Space-Time Network-Based Modeling Framework for Dynamic Unmanned Aerial Vehicle Routing in Traffic Incident Monitoring Applications.

    Science.gov (United States)

    Zhang, Jisheng; Jia, Limin; Niu, Shuyun; Zhang, Fan; Tong, Lu; Zhou, Xuesong

    2015-01-01

    It is essential for transportation management centers to equip and manage a network of fixed and mobile sensors in order to quickly detect traffic incidents and further monitor the related impact areas, especially for high-impact accidents with dramatic traffic congestion propagation. As emerging small Unmanned Aerial Vehicles (UAVs) start to have a more flexible regulation environment, it is critically important to fully explore the potential for of using UAVs for monitoring recurring and non-recurring traffic conditions and special events on transportation networks. This paper presents a space-time network- based modeling framework for integrated fixed and mobile sensor networks, in order to provide a rapid and systematic road traffic monitoring mechanism. By constructing a discretized space-time network to characterize not only the speed for UAVs but also the time-sensitive impact areas of traffic congestion, we formulate the problem as a linear integer programming model to minimize the detection delay cost and operational cost, subject to feasible flying route constraints. A Lagrangian relaxation solution framework is developed to decompose the original complex problem into a series of computationally efficient time-dependent and least cost path finding sub-problems. Several examples are used to demonstrate the results of proposed models in UAVs' route planning for small and medium-scale networks. PMID:26076404

  1. A fast U-D factorization-based learning algorithm with applications to nonlinear system modeling and identification.

    Science.gov (United States)

    Zhang, Y; Li, X R

    1999-01-01

    A fast learning algorithm for training multilayer feedforward neural networks (FNN's) by using a fading memory extended Kalman filter (FMEKF) is presented first, along with a technique using a self-adjusting time-varying forgetting factor. Then a U-D factorization-based FMEKF is proposed to further improve the learning rate and accuracy of the FNN. In comparison with the backpropagation (BP) and existing EKF-based learning algorithms, the proposed U-D factorization-based FMEKF algorithm provides much more accurate learning results, using fewer hidden nodes. It has improved convergence rate and numerical stability (robustness). In addition, it is less sensitive to start-up parameters (e.g., initial weights and covariance matrix) and the randomness in the observed data. It also has good generalization ability and needs less training time to achieve a specified learning accuracy. Simulation results in modeling and identification of nonlinear dynamic systems are given to show the effectiveness and efficiency of the proposed algorithm. PMID:18252590

  2. Asymptotics-based CI models for atoms:Properties, exact solution of a minimal model for Li to Ne, and application to atomic spectra

    OpenAIRE

    Friesecke, G.; Goddard, B.D.

    2009-01-01

    Configuration-interaction (CI) models are approximations to the electronic Schrödinger equation which are widely used for numerical electronic structure calculations in quantum chemistry. Based on our recent closed-form asymptotic results for the full atomic Schrödinger equation in the limit of fixed electron number and large nuclear charge [SIAM J. Math. Anal., 41 (2009), pp. 631-664], we introduce a class of CI models for atoms which reproduce, at fixed finite model dimension, the correct S...

  3. Disaggregation of nation-wide dynamic population exposure estimates in The Netherlands: Applications of activity-based transport models

    Science.gov (United States)

    Beckx, Carolien; Int Panis, Luc; Uljee, Inge; Arentze, Theo; Janssens, Davy; Wets, Geert

    Traditional exposure studies that link concentrations with population data do not always take into account the temporal and spatial variations in both concentrations and population density. In this paper we present an integrated model chain for the determination of nation-wide exposure estimates that incorporates temporally and spatially resolved information about people's location and activities (obtained from an activity-based transport model) and about ambient pollutant concentrations (obtained from a dispersion model). To the best of our knowledge, it is the first time that such an integrated exercise was successfully carried out in a fully operational modus for all models under consideration. The evaluation of population level exposure in The Netherlands to NO 2 at different time-periods, locations, for different subpopulations (gender, socio-economic status) and during different activities (residential, work, transport, shopping) is chosen as a case-study to point out the new features of this methodology. Results demonstrate that, by neglecting people's travel behaviour, total average exposure to NO 2 will be underestimated by 4% and hourly exposure results can be underestimated by more than 30%. A more detailed exposure analysis reveals the intra-day variations in exposure estimates and the presence of large exposure differences between different activities (traffic > work > shopping > home) and between subpopulations (men > women, low socio-economic class > high socio-economic class). This kind of exposure analysis, disaggregated by activities or by subpopulations, per time of day, provides useful insight and information for scientific and policy purposes. It demonstrates that policy measures, aimed at reducing the overall (average) exposure concentration of the population may impact in a different way depending on the time of day or the subgroup considered. From a scientific point of view, this new approach can be used to reduce exposure misclassification.

  4. Methane emissions from floodplains in the Amazon Basin: towards a process-based model for global applications

    Directory of Open Access Journals (Sweden)

    B. Ringeval

    2013-10-01

    Full Text Available Tropical wetlands are estimated to represent about 50% of the natural wetland emissions and explain a large fraction of the observed CH4 variability on time scales ranging from glacial-interglacial cycles to the currently observed year-to-year variability. Despite their importance, however, tropical wetlands are poorly represented in global models aiming to predict global CH4 emissions. This study documents the first regional-scale, process-based model of CH4 emissions from tropical floodplains. The LPX-Bern Dynamic Global Vegetation Model (LPX hereafter was modified to represent floodplain hydrology, vegetation and associated CH4 emissions. The extent of tropical floodplains was prescribed using output from the spatially-explicit hydrology model PCR-GLOBWB. We introduced new Plant Functional Types (PFTs that explicitly represent floodplain vegetation. The PFT parameterizations were evaluated against available remote sensing datasets (GLC2000 land cover and MODIS Net Primary Productivity. Simulated CH4 flux densities were evaluated against field observations and regional flux inventories. Simulated CH4 emissions at Amazon Basin scale were compared to model simulations performed in the WETCHIMP intercomparison project. We found that LPX simulated CH4 flux densities are in reasonable agreement with observations at the field scale but with a~tendency to overestimate the flux observed at specific sites. In addition, the model did not reproduce between-site variations or between-year variations within a site. Unfortunately, site informations are too limited to attest or disprove some model features. At the Amazon Basin scale, our results underline the large uncertainty in the magnitude of wetland CH4 emissions. In particular, uncertainties in floodplain extent (i.e., difference between GLC2000 and PCR-GLOBWB output modulate the simulated emissions by a factor of about 2. Our best estimates, using PCR-GLOBWB in combination with GLC2000, lead to

  5. A Framework for Next Generation Mobile and Wireless Networks Application Development using Hybrid Component Based Development Model

    OpenAIRE

    Barnawi, Ahmed; Qureshi, M. Rizwan Jameel; Khan, Asif Irshad

    2012-01-01

    The IP Multimedia Subsystems (IMS) that features in Next Generation Networks (NGN) offers the application developer (third party) abilities to map out applications over mobile telecommunication infrastructure. The IMS comes about with APIs useful for mobile application developers to create applications to meet end-users' demands and comply with the provider's infrastructure set up at the same time. Session Initiation Protocol (SIP) is a signaling protocol for this architecture. It is used for...

  6. Simulation of turbulent supersonic separated base flows using enhanced turbulence modeling techniques with application to an X-33 aerospike rocket nozzle system

    Science.gov (United States)

    Papp, John Laszlo

    2000-10-01

    The successful application of CFD and turbulence modeling methods to an aerospike nozzle system first involves the successful simulation of its key flow components. This report addresses the task using the Chien low-Re k-epsilon and the Yakhot et al. high-Re RNG k-epsilon turbulence models. An improved implicit axis of symmetry boundary condition is also developed to increase stability and lower artificial dissipation. Grid adaptation through the SAGE post-processing package is used throughout the study. The RNG model, after low-Re modifications, and the Chien low-Re k-epsilon model are applied to the supersonic axisymmetric base flow problem. Both models predict a peak recirculation velocity almost twice as large as experiment. The RNG model predicts a flatter base pressure and lower recirculation velocity more consistent with experimental data using less grid points than a comparable Chien model solution. The turbulent quantities predicted by both models are typical of other numerical results and generally under predict peak values obtained in experiment suggesting that too little turbulent eddy viscosity is produced. After several test cases, the full 3-D aerospike nozzle is simulated using both the Chien and modified RNG low-Re models. The Chien model outperforms the RNG model in all circumstances. The surface pressure predicted by the Chien model along the nozzle center-plane is very near experiment while mid-plane results are not as close but useful for design purposes. The lack of a thick boundary layer along the nozzle surface in RNG simulations is the cause of poor surface pressure comparisons. Although initial base flow comparisons between the model predictions and experiment are poor, the profiles are relatively flat. To accelerate the progress to a steady-state solution, a process involving the artificial lowering of the base pressure and subsequent iteration to a new steady state is undertaken. After several of these steps, the resulting steady

  7. Small signal model parameters analysis of GaN and GaAs based HEMTs over temperature for microwave applications

    Science.gov (United States)

    Alim, Mohammad A.; Rezazadeh, Ali A.; Gaquiere, Christophe

    2016-05-01

    Thermal and small-signal model parameters analysis have been carried out on 0.5 μm × (2 × 100 μm) AlGaAs/GaAs HEMT grown on semi-insulating GaAs substrate and 0.25 μm × (2 × 100 μm) AlGaN/GaN HEMT grown on SiC substrate. Two different technologies are investigated in order to establish a detailed understanding of their capabilities in terms of frequency and temperature using on-wafer S-parameter measurement over the temperature range from -40 to 150 °C up to 50 GHz. The equivalent circuit parameters as well as their temperature-dependent behavior of the two technologies were analyzed and discussed for the first time. The principle elevation or degradation of transistor parameters with temperature demonstrates the great potential of GaN device for high frequency and high temperature applications. The result provides some valuable insights for future design optimizations of advanced GaN and a comparison of this with the GaAs technology.

  8. A solely radiance-based spectral anisotropic distribution model and its application in deriving clear-sky spectral fluxes

    Science.gov (United States)

    Song, L.; Huang, X.

    2011-12-01

    Anisotropic distribution model (ADM) plays a uniquely central role in converting broadband radiance measurement to broadband flux. Scene type classifications are usually needed for such ADM and such classifications are usually done with auxiliary measurements and information since broadband radiance does not contain detailed information about temperature, humidity, and clouds. Recently Huang et al. (2008 and 2010) has developed spectral ADM based on such scene type classifications and successfully derived spectral flux from spectral radiance measurement. Unlike broadband radiances, the spectrally resolved radiances indeed contain rich information about temperature, humidity, and clouds. Therefore, it is meaningful to explore whether it is possible to develop scene-type classification solely based on spectral radiance and consequently to construct spectral ADM solely base on radiances measurement. Using AIRS spectrum as an example, here we develop a clear-sky scene classification algorithm solely based on AIRS radiances. The definitions of scene types are similar to those of clear-sky scene types used in CERES SSF algorithm, which are discrete intervals based on surface skin temperature, lapse rate (temperature change of the first 300 mb above the surface), and the total precipitable water (TPW). Brightness temperature of AIRS channel at 963.8 cm-1 are used for determine corresponding discrete intervals of surface skin temperature. This channel is also used in conjunction with a channel at 748.6 cm-1 for categorizing the lapse rate. Given the slow varying of water vapor continuum in the window region and the dominant weight of lower tropospheric humidity in TPW, a double-differential technique is used to categorize the TPW. By choosing two pairs of AIRS channels with similar frequency intervals, the technique can classify the TPW without any a priori information about continuum absorption since double differencing largely remove the slow-varying continuum

  9. Model-based fault detection and isolation for intermittently active faults with application to motion-based thruster fault detection and isolation for spacecraft

    Science.gov (United States)

    Wilson, Edward (Inventor)

    2008-01-01

    The present invention is a method for detecting and isolating fault modes in a system having a model describing its behavior and regularly sampled measurements. The models are used to calculate past and present deviations from measurements that would result with no faults present, as well as with one or more potential fault modes present. Algorithms that calculate and store these deviations, along with memory of when said faults, if present, would have an effect on the said actual measurements, are used to detect when a fault is present. Related algorithms are used to exonerate false fault modes and finally to isolate the true fault mode. This invention is presented with application to detection and isolation of thruster faults for a thruster-controlled spacecraft. As a supporting aspect of the invention, a novel, effective, and efficient filtering method for estimating the derivative of a noisy signal is presented.

  10. Dual Security Testing Model for Web Applications

    Directory of Open Access Journals (Sweden)

    Singh Garima

    2016-02-01

    Full Text Available In recent years, web applications have evolved from small websites into large multi-tiered applications. The quality of web applications depends on the richness of contents, well structured navigation and most importantly its security. Web application testing is a new field of research so as to ensure the consistency and quality of web applications. In the last ten years there have been different approaches. Models have been developed for testing web applications but only a few focused on content testing, a few on navigation testing and a very few on security testing of web applications. There is a need to test content, navigation and security of an application in one go. The objective of this paper is to propose Dual Security Testing Model to test the security of web applications using UML modeling technique which includes web socket interface. In this research paper we have described how our security testing model is implemented using activity diagram, activity graph and based on this how test cases is generated.

  11. Web-based applications for virtual laboratories

    NARCIS (Netherlands)

    Bier, H.H.

    2011-01-01

    Web-based applications for academic education facilitate, usually, exchange of multimedia files, while design-oriented domains such as architectural and urban design require additional support in collaborative real-time drafting and modeling. In this context, multi-user interactive interfaces employ

  12. Toughness Scaling Model Applications

    Czech Academy of Sciences Publication Activity Database

    Dlouhý, Ivo; Kozák, Vladislav; Holzmann, Miloslav

    78. Dordrecht : Kluwer Academic Publishers, 2002 - (Dlouhý, I.), s. 195-212 - (NATO Science Series. Mathematics, Physics and Chemistry. 2) R&D Projects: GA AV ČR IAA2041003; GA MŠk ME 303 Institutional research plan: CEZ:AV0Z2041904 Keywords : Fracture toughness transferability * pre cracked Charpyspecimen * toughness scaling models Subject RIV: JL - Materials Fatigue, Friction Mechanics

  13. Sociable interface-based model checking for Web applications%基于交际接口的Web应用模型检验

    Institute of Scientific and Technical Information of China (English)

    李决龙; 李亮; 邢建春; 杨启亮

    2011-01-01

    为了验证Web应用的质量,首次采用了基于交际接口及其工具TICC的建筑智能化系统Web应用验证方法,通过一个简单的能源管理Web应用系统实例说明了整个建模、构件模块组合验证和系统性质验证过程.结果表明验证能够顺利实现,因而该方法是一种合适的Web应用验证方法.%In order to verify Web applications' quality, the paper firstly adopted the methodology based on sociable interface and its tool TICC to check Web applications in the intelligent building systems, used a simple case of energy sources management Web application system to illustrate the whole process of modeling, component composing verification and characteristic model checking.The result shows that verification is done successfully, so it is an appropriate verification method for Web applications.

  14. Variability of tsunami inundation footprints considering stochastic scenarios based on a single rupture model: Application to the 2011 Tohoku earthquake

    Science.gov (United States)

    Goda, Katsuichiro; Yasuda, Tomohiro; Mori, Nobuhito; Mai, P. Martin

    2015-06-01

    The sensitivity and variability of spatial tsunami inundation footprints in coastal cities and towns due to a megathrust subduction earthquake in the Tohoku region of Japan are investigated by considering different fault geometry and slip distributions. Stochastic tsunami scenarios are generated based on the spectral analysis and synthesis method with regards to an inverted source model. To assess spatial inundation processes accurately, tsunami modeling is conducted using bathymetry and elevation data with 50 m grid resolutions. Using the developed methodology for assessing variability of tsunami hazard estimates, stochastic inundation depth maps can be generated for local coastal communities. These maps are important for improving disaster preparedness by understanding the consequences of different situations/conditions, and by communicating uncertainty associated with hazard predictions. The analysis indicates that the sensitivity of inundation areas to the geometrical parameters (i.e., top-edge depth, strike, and dip) depends on the tsunami source characteristics and the site location, and is therefore complex and highly nonlinear. The variability assessment of inundation footprints indicates significant influence of slip distributions. In particular, topographical features of the region, such as ria coast and near-shore plain, have major influence on the tsunami inundation footprints.

  15. Variability of tsunami inundation footprints considering stochastic scenarios based on a single rupture model: Application to the 2011 Tohoku earthquake

    KAUST Repository

    Goda, Katsuichiro

    2015-06-30

    The sensitivity and variability of spatial tsunami inundation footprints in coastal cities and towns due to a megathrust subduction earthquake in the Tohoku region of Japan are investigated by considering different fault geometry and slip distributions. Stochastic tsunami scenarios are generated based on the spectral analysis and synthesis method with regards to an inverted source model. To assess spatial inundation processes accurately, tsunami modeling is conducted using bathymetry and elevation data with 50 m grid resolutions. Using the developed methodology for assessing variability of tsunami hazard estimates, stochastic inundation depth maps can be generated for local coastal communities. These maps are important for improving disaster preparedness by understanding the consequences of different situations/conditions, and by communicating uncertainty associated with hazard predictions. The analysis indicates that the sensitivity of inundation areas to the geometrical parameters (i.e., top-edge depth, strike, and dip) depends on the tsunami source characteristics and the site location, and is therefore complex and highly nonlinear. The variability assessment of inundation footprints indicates significant influence of slip distributions. In particular, topographical features of the region, such as ria coast and near-shore plain, have major influence on the tsunami inundation footprints.

  16. Modeling techniques and processes control application based on Neural Networks with on-line adjustment using Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    R. F. Marcolla

    2009-03-01

    Full Text Available In this work a strategy is presented for the temperature control of the polymerization reaction of styrene in suspension in batch. A three-layer feed forward Artificial Neural Network was trained in an off-line way starting from a removed group of patterns of the experimental system and applied in the recurrent form (RNN to a Predictive Controller based on a Nonlinear Model (NMPC. This controller presented very superior results to the classic controller PID in the maintenance of the temperature. Still to improve the performance of the model used by NMPC (RNN that can present differences in relation to the system due to the dead time involved in the control actions, nonlinear characteristic of the system and variable dynamics; an on-line adjustment methodology of the parameters of the exit layer of the Network is implemented, presenting superior results and treating the difficulties satisfactorily in the temperature control. All the presented results are obtained for a real system.

  17. Application of a Physiologically Based Pharmacokinetic Model to Study Theophylline Metabolism and Its Interactions With Ciprofloxacin and Caffeine.

    Science.gov (United States)

    Navid, A; Ng, D M; Wong, S E; Lightstone, F C

    2016-02-01

    Theophylline is a commonly used bronchodilator. However, due to its narrow therapeutic range, moderate elevation of serum concentration can result in adverse drug reactions (ADRs). ADRs occur because of interhuman pharmacokinetic variability and interactions with coprescribed medicines. We developed a physiologically based pharmacokinetic (PBPK) model of theophylline, caffeine, and ciprofloxacin metabolisms to: examine theophylline pharmacokinetic variability, and predict population-level outcomes of drug-drug interactions (DDIs). A simulation-based equation for personalized dosing of theophylline was derived. Simulations of DDI show that calculated personalized doses are safe even after cotreatment with large doses of strong inhibitors. Simulations of adult populations indicate that the elderly are most susceptible to ADRs stemming from theophylline-ciprofloxacin and theophylline-caffeine interactions. Females, especially Asians, due to their smaller average size, are more susceptible to DDI-induced ADRs following typical dosing practices. Our simulations also show that the higher adipose and lower muscle fractions in females significantly alter the pharmacokinetics of theophylline or ciprofloxacin. PMID:26933518

  18. Application of a rule-based model to estimate mercury exchange for three background biomes in the continental United States

    Science.gov (United States)

    Hartman, J.S.; Weisberg, P.J.; Pillai, R.; Ericksen, J.A.; Kuiken, T.; Lindberg, S.E.; Zhang, H.; Rytuba, J.J.; Gustin, M.S.

    2009-01-01

    Ecosystems that have low mercury (Hg) concentrations (i.e., not enriched or impactedbygeologic or anthropogenic processes) cover most of the terrestrial surface area of the earth yet their role as a net source or sink for atmospheric Hg is uncertain. Here we use empirical data to develop a rule-based model implemented within a geographic information system framework to estimate the spatial and temporal patterns of Hg flux for semiarid deserts, grasslands, and deciduous forests representing 45% of the continental United States. This exercise provides an indication of whether these ecosystems are a net source or sink for atmospheric Hg as well as a basis for recommendation of data to collect in future field sampling campaigns. Results indicated that soil alone was a small net source of atmospheric Hg and that emitted Hg could be accounted for based on Hg input by wet deposition. When foliar assimilation and wet deposition are added to the area estimate of soil Hg flux these biomes are a sink for atmospheric Hg. ?? 2009 American Chemical Society.

  19. Actuarial applications of financial models.

    OpenAIRE

    Goovaerts, Marc; Dhaene, Jan

    1996-01-01

    In the present contribution we indicate the type of situations seen from an insurance point of view, in which financial models serve as a basis for providing solutions to practical problems . In addition, some of the essential differences in the basic assumptions underlying financial models and actuarial applications are given.

  20. A Framework for Next Generation Mobile and Wireless Networks Application Development using Hybrid Component Based Development Model

    CERN Document Server

    Barnawi, Ahmed; Khan, Asif Irshad

    2012-01-01

    The IP Multimedia Subsystems (IMS) that features in Next Generation Networks (NGN) offers the application developer (third party) abilities to map out applications over mobile telecommunication infrastructure. The IMS comes about with APIs useful for mobile application developers to create applications to meet end-users' demands and comply with the provider's infrastructure set up at the same time. Session Initiation Protocol (SIP) is a signaling protocol for this architecture. It is used for establishing sessions in IP network, making it an ideal candidate for supporting terminal mobility in to deliver the services with improved Quality of Services (QOS). The realization of IMS's virtues as far as software design is concerned is faced by lack of standardizations and methodologies throughout application development process. In this paper, we report on progress on ongoing research by our group toward putting together a platform as a testbed used for NGN application development. We examine a novel component bas...

  1. Cell-Based Biosensors Principles and Applications

    CERN Document Server

    Wang, Ping

    2009-01-01

    Written by recognized experts the field, this leading-edge resource is the first book to systematically introduce the concept, technology, and development of cell-based biosensors. You find details on the latest cell-based biosensor models and novel micro-structure biosensor techniques. Taking an interdisciplinary approach, this unique volume presents the latest innovative applications of cell-based biosensors in a variety of biomedical fields. The book also explores future trends of cell-based biosensors, including integrated chips, nanotechnology and microfluidics. Over 140 illustrations hel

  2. Hydraulic model for multi-sources reclaimed water pipe network based on EPANET and its applications in Beijing, China

    Institute of Scientific and Technical Information of China (English)

    Haifeng JIA; Wei WEI; Kunlun XIN

    2008-01-01

    Water shortage is one of the major water related problems for many cities in the world. The planning for utilization of reclaimed water has been or would be drafted in these cities. For using the reclaimed water soundly, Beijing planned to build a large scale reclaimed water pipe networks with multi-sources. In order to support the plan, the integrated hydraulic model of planning pipe network was developed based on EPANET supported by geographic information system (GIS). The complicated pipe network was divided into four weak conjunction subzones according to the distribution of reclaimed water plants and the elevation. It could provide a better solution for the problem of overhigh pressure in several regions of the network. Through the scenarios analy-sis in different subzones, some of the initial diameter of pipes in the network was adjusted. At last the pipe network planning scheme of reclaimed water was proposed. The proposed planning scheme could reach the balances between reclaimed water requirements and reclaimed water supplies, and provided a scientific basis for the reclaimed water utilization in Beijing. Now the scheme had been adopted by Beijing municipal government.

  3. Location Based Services and Applications

    Directory of Open Access Journals (Sweden)

    Elenis Gorrita Michel

    2012-05-01

    Full Text Available Location Based Services (LBS continue to grow in popularity, effectiveness and reliability, to the extent that applications are designed and implemented taking into account the facilities of the user location information. In this work, some of the main applications are addressed, in order to make an assessment of the current importance of the LBS, as a branch of technology in full swing. In addition, the main techniques for location estimation are studied, essential information to the LBS. Because of this it is a highly topical issue, the ongoing works and researches are also discussed.

  4. From situation modelling to a distributed rule-based platform for situation awareness: an ontological framework for disaster management applications

    NARCIS (Netherlands)

    Moreira, João Luiz Rebelo

    2015-01-01

    Situation-aware (SA) applications are particularly useful for disaster management. The complex nature of emergency scenarios presents challenges to the development of collaborative and distributed SA solutions. These challenges concern the whole lifecycle, from specification to implementation phases

  5. Applicability of heat and gas trans-port models in biocover design based on a case study from Denmark

    DEFF Research Database (Denmark)

    Nielsen, A. A. F.; Binning, Philip John; Kjeldsen, Peter

    2015-01-01

    Biocovers — layers of mature compost — can oxidise a considerable amount of methane emitted from landfi Different factors can affect oxidation, particularly tempera- ture. For better understanding of the processes and for future biocover designs, two models (analytic and numerical) were developed....... Both models used the heat equation for heat transfer, and the numerical model used advection-diffusion model with dual Monod kinetics for gas transport. The results were validated with data from a Danish landfi The models correlated well with the observed data: the coefficient of determination (R2) was...... 0.95 for the analytic model and 0.91 for the numerical model. The models can be used for different design scenarios (e.g. varying methane infl thickness or start of operation), and can also help understand the processes that take place in the system, e.g. how oxygen penetration depends on ambient...

  6. Aviation Safety Modeling and Simulation (ASMM) Propulsion Fleet Modeling: A Tool for Semi-Automatic Construction of CORBA-based Applications from Legacy Fortran Programs

    Science.gov (United States)

    Sang, Janche

    2003-01-01

    Within NASA's Aviation Safety Program, NASA GRC participates in the Modeling and Simulation Project called ASMM. NASA GRC s focus is to characterize the propulsion systems performance from a fleet management and maintenance perspective by modeling and through simulation predict the characteristics of two classes of commercial engines (CFM56 and GE90). In prior years, the High Performance Computing and Communication (HPCC) program funded, NASA Glenn in developing a large scale, detailed simulations for the analysis and design of aircraft engines called the Numerical Propulsion System Simulation (NPSS). Three major aspects of this modeling included the integration of different engine components, coupling of multiple disciplines, and engine component zooming at appropriate level fidelity, require relatively tight coupling of different analysis codes. Most of these codes in aerodynamics and solid mechanics are written in Fortran. Refitting these legacy Fortran codes with distributed objects can increase these codes reusability. Aviation Safety s modeling and simulation use in characterizing fleet management has similar needs. The modeling and simulation of these propulsion systems use existing Fortran and C codes that are instrumental in determining the performance of the fleet. The research centers on building a CORBA-based development environment for programmers to easily wrap and couple legacy Fortran codes. This environment consists of a C++ wrapper library to hide the details of CORBA and an efficient remote variable scheme to facilitate data exchange between the client and the server model. Additionally, a Web Service model should also be constructed for evaluation of this technology s use over the next two- three years.

  7. Survival analysis models and applications

    CERN Document Server

    Liu, Xian

    2012-01-01

    Survival analysis concerns sequential occurrences of events governed by probabilistic laws.  Recent decades have witnessed many applications of survival analysis in various disciplines. This book introduces both classic survival models and theories along with newly developed techniques. Readers will learn how to perform analysis of survival data by following numerous empirical illustrations in SAS. Survival Analysis: Models and Applications: Presents basic techniques before leading onto some of the most advanced topics in survival analysis.Assumes only a minimal knowledge of SAS whilst enablin

  8. Application of flood risk modelling in a web-based geospatial decision support tool for coastal adaptation to climate change

    Science.gov (United States)

    Knight, P. J.; Prime, T.; Brown, J. M.; Morrissey, K.; Plater, A. J.

    2015-07-01

    A pressing problem facing coastal decision makers is the conversion of "high-level" but plausible climate change assessments into an effective basis for climate change adaptation at the local scale. Here, we describe a web-based, geospatial decision support tool (DST) that provides an assessment of the potential flood risk for populated coastal lowlands arising from future sea-level rise, coastal storms, and high river flows. This DST has been developed to support operational and strategic decision making by enabling the user to explore the flood hazard from extreme events, changes in the extent of the flood-prone areas with sea-level rise, and thresholds of sea-level rise where current policy and resource options are no longer viable. The DST is built in an open-source GIS that uses freely available geospatial data. Flood risk assessments from a combination of LISFLOOD-FP and SWAB (Shallow Water And Boussinesq) models are embedded within the tool; the user interface enables interrogation of different combinations of coastal and river events under rising-sea-level scenarios. Users can readily vary the input parameters (sea level, storms, wave height and river flow) relative to the present-day topography and infrastructure to identify combinations where significant regime shifts or "tipping points" occur. Two case studies demonstrate the attributes of the DST with respect to the wider coastal community and the UK energy sector. Examples report on the assets at risk and illustrate the extent of flooding in relation to infrastructure access. This informs an economic assessment of potential losses due to climate change and thus provides local authorities and energy operators with essential information on the feasibility of investment for building resilience into vulnerable components of their area of responsibility.

  9. Application of flood risk modelling in a web-based geospatial decision support tool for coastal adaptation to climate change

    Directory of Open Access Journals (Sweden)

    P. J. Knight

    2015-02-01

    Full Text Available A pressing problem facing coastal decision makers is the conversion of "high level" but plausible climate change assessments into an effective basis for climate change adaptation at the local scale. Here, we describe a web-based, geospatial decision-support tool (DST that provides an assessment of the potential flood risk for populated coastal lowlands arising from future sea-level rise, coastal storms and high river flows. This DST has been developed to support operational and strategic decision making by enabling the user to explore the flood hazard from extreme events, changes in the extent of the flood-prone areas with sea-level rise, and thresholds of sea-level rise where current policy and resource options are no longer viable. The DST is built in an open source GIS that uses freely available geospatial data. Flood risk assessments from a combination of LISFLOOD-FP and SWAB models are embedded within the tool; the user interface enables interrogation of different combinations of coastal and river events under rising sea-level scenarios. Users can readily vary the input parameters (sea level, storms, wave height and river flow relative to the present-day topography and infrastructure to identify combinations where significant regime shifts or "tipping points" occur. Two case studies are used to demonstrate the attributes of the DST with respect to the wider coastal community and the UK energy sector. Examples report on the assets at risk and illustrate the extent of flooding in relation to infrastructure access. This informs an economic assessment of potential losses due to climate change and thus provides local authorities and energy operators with essential information on the feasibility of investment for building resilience into vulnerable components of their area of responsibility.

  10. Application of wavelet neural network model based on genetic algorithm in the prediction of high-speed railway settlement

    Science.gov (United States)

    Tang, Shihua; Li, Feida; Liu, Yintao; Lan, Lan; Zhou, Conglin; Huang, Qing

    2015-12-01

    With the advantage of high speed, big transport capacity, low energy consumption, good economic benefits and so on, high-speed railway is becoming more and more popular all over the world. It can reach 350 kilometers per hour, which requires high security performances. So research on the prediction of high-speed railway settlement that as one of the important factors affecting the safety of high-speed railway becomes particularly important. This paper takes advantage of genetic algorithms to seek all the data in order to calculate the best result and combines the advantage of strong learning ability and high accuracy of wavelet neural network, then build the model of genetic wavelet neural network for the prediction of high-speed railway settlement. By the experiment of back propagation neural network, wavelet neural network and genetic wavelet neural network, it shows that the absolute value of residual errors in the prediction of high-speed railway settlement based on genetic algorithm is the smallest, which proves that genetic wavelet neural network is better than the other two methods. The correlation coefficient of predicted and observed value is 99.9%. Furthermore, the maximum absolute value of residual error, minimum absolute value of residual error-mean value of relative error and value of root mean squared error(RMSE) that predicted by genetic wavelet neural network are all smaller than the other two methods'. The genetic wavelet neural network in the prediction of high-speed railway settlement is more stable in terms of stability and more accurate in the perspective of accuracy.

  11. The Development of Dynamic Brand Equity Chase Model and Its Application to Digital Industry Based on Scanner Data

    Directory of Open Access Journals (Sweden)

    Nam Yongsik

    2009-12-01

    Full Text Available The purpose of this research is to develop a comprehensive modeling for measuring dynamics of brand power. We define brand power as brand specific coefficients to yield the sales volume for each period. The modeling consists of multinomial log it model for eachproduct category, the brand-specific coefficients, mixture modeling and fuzzy clustering algorithm. We apply our modeling to TV scanner data in Tianjin China. The results show 5 brands have 12 to 23 times change on their brand power in a year. The lasting time of brandpower spreads from 1 week to 12 weeks.

  12. Optimal pricing decision model based on activity-based costing

    Institute of Scientific and Technical Information of China (English)

    王福胜; 常庆芳

    2003-01-01

    In order to find out the applicability of the optimal pricing decision model based on conventional costbehavior model after activity-based costing has given strong shock to the conventional cost behavior model andits assumptions, detailed analyses have been made using the activity-based cost behavior and cost-volume-profitanalysis model, and it is concluded from these analyses that the theory behind the construction of optimal pri-cing decision model is still tenable under activity-based costing, but the conventional optimal pricing decisionmodel must be modified as appropriate to the activity-based costing based cost behavior model and cost-volume-profit analysis model, and an optimal pricing decision model is really a product pricing decision model construc-ted by following the economic principle of maximizing profit.

  13. Research and Application of Role-Based Access Control Model in Web Application System%Web应用系统中RBAC模型的研究与实现

    Institute of Scientific and Technical Information of China (English)

    黄秀文

    2015-01-01

    Access control is the main strategy of security and protection in Web system, the traditional access control can not meet the needs of the growing security. With using the role based access control (RBAC) model and introducing the concept of the role in the web system, the user is mapped to a role in an organization, access to the corresponding role authorization, access authorization and control according to the user's role in an organization, so as to improve the web system flexibility and security permissions and access control.%访问控制是Web系统中安全防范和保护的主要策略,传统的访问控制已不能满足日益增长的安全性需求。本文在web应用系统中,使用基于角色的访问控制(RBAC)模型,通过引入角色的概念,将用户映射为在一个组织中的某种角色,将访问权限授权给相应的角色,根据用户在组织内所处的角色进行访问授权与控制,从而提高了在web系统中权限分配和访问控制的灵活性与安全性。

  14. Application of a model based on a pair of Laplace transforms for standard low-energy X-ray beams spectral reconstruction

    International Nuclear Information System (INIS)

    The direct measurement of the spectrum of an X-ray beam by some spectroscopic method is relatively difficult and expensive. Spectra can be alternatively derived by an indirect method from measurements of transmission curve of the X-ray beam and the use of Laplace transforms. The objective of this work was the application of an indirect method that use a spectral model based on a pair of Laplace transforms to reconstruct experimental published spectra of standard low energy X-ray beams at radiation protection level and determine the mean photon energy from the reconstructed spectra for radiation quality specification. The spectral model was applied using calculated transmission curves and the reconstructed spectra provided a coarse approximation to experimental data. Even though, the mean photon energy of the X-ray beams determined from these reconstructed spectra present a satisfactory result showing the value of the analysis of transmission curves for the X-ray beam quality specification. (author)

  15. Research protocol: EB-GIS4HEALTH UK – foundation evidence base and ontology-based framework of modular, reusable models for UK/NHS health and healthcare GIS applications

    Directory of Open Access Journals (Sweden)

    Boulos Maged

    2005-01-01

    Full Text Available Abstract EB-GIS4HEALTH UK aims at building a UK-oriented foundation evidence base and modular conceptual models for GIS applications and programmes in health and healthcare to improve the currently poor GIS state of affairs within the NHS; help the NHS understand and harness the importance of spatial information in the health sector in order to better respond to national health plans, priorities, and requirements; and also foster the much-needed NHS-academia GIS collaboration. The project will focus on diabetes and dental care, which together account for about 11% of the annual NHS budget, and are thus important topics where GIS can help optimising resource utilisation and outcomes. Virtual e-focus groups will ensure all UK/NHS health GIS stakeholders are represented. The models will be built using Protégé ontology editor http://protege.stanford.edu/ based on the best evidence pooled in the project's evidence base (from critical literature reviews and e-focus groups. We will disseminate our evidence base, GIS models, and documentation through the project's Web server. The models will be human-readable in different ways to inform NHS GIS implementers, and it will be possible to also use them to generate the necessary template databases (and even to develop "intelligent" health GIS solutions using software agents for running the modelled applications. Our products and experience in this project will be transferable to address other national health topics based on the same principles. Our ultimate goal is to provide the NHS with practical, vendor-neutral, modular workflow models, and ready-to-use, evidence-based frameworks for developing successful GIS business plans and implementing GIS to address various health issues. NHS organisations adopting such frameworks will achieve a common understanding of spatial data and processes, which will enable them to efficiently and effectively share, compare, and integrate their data silos and results for

  16. CALCINATION AND SINTERING MODELS FOR APPLICATION TO HIGH-TEMPERATURE, SHORT-TIME SULFATION OF CALCIUM-BASED SORBENTS

    Science.gov (United States)

    To simulate the staged availability of transient high surface area CaO observed in high-temperature flow-reactor data, the rate of calcination of CaCO3 or Ca(OH)2 is described by an empirical modification of the shrinking-core model. The physical model depicts particle decomposi...

  17. Design of pulsed perforated-plate columns for industrial scale mass transfer applications - present experience and the need for a model based approach

    International Nuclear Information System (INIS)

    Mass transfer is a vital unit operation in the processing of spent nuclear fuel in the backend of closed fuel cycle and Pulsed perforated plate extraction columns are used as mass transfer device for more than five decades. The pulsed perforated plate column is an agitated differential contactor, which has wide applicability due to its simplicity, high mass transfer efficiency, high through put, suitability for maintenance free remote operation, ease of cleaning/decontamination and cost effectiveness. Design of pulsed columns are based on a model proposed to describe the hydrodynamics and mass transfer. In equilibrium stage model, the HETS values are obtained from pilot plant experiments and then scaled empirically to design columns for industrial application. The dispersion model accounts for mass transfer kinetics and back-mixing. The drop population balance model can describe complex hydrodynamics of dispersed phase, that is, drop formation, break-up and drop-to-drop interactions. In recent years, significant progress has been made to model pulsed columns using CFD, which provides complete mathematical description of hydrodynamics in terms of spatial distribution of flow fields and 3D visualization. Under the condition of pulsation, the poly-dispersed nature of turbulent droplet swarm renders modeling difficult. In the absence of industry acceptance of proposed models, the conventional chemical engineering practice is to use HETS-NTS concept or HTU-NTU approach to design extraction columns. The practicability of HTU-NTU approach has some limitations due to the lack of experimental data on individual film mass transfer coefficients. Presently, the HETS-NTS concept has been used for designing the columns, which has given satisfactory performance. The design objective is mainly to arrive at the diameter and height of the mass transfer section for a specific plate geometry, fluid properties and pulsing condition to meet the intended throughput (capacity) and mass

  18. Behavior Modeling -- Foundations and Applications

    DEFF Research Database (Denmark)

    This book constitutes revised selected papers from the six International Workshops on Behavior Modelling - Foundations and Applications, BM-FA, which took place annually between 2009 and 2014. The 9 papers presented in this volume were carefully reviewed and selected from a total of 58 papers...

  19. Applications of Continuum Shell Model

    OpenAIRE

    Volya, Alexander

    2006-01-01

    The nuclear many-body problem at the limits of stability is considered in the framework of the Continuum Shell Model that allows a unified description of intrinsic structure and reactions. Technical details behind the method are highlighted and practical applications combining the reaction and structure pictures are presented.

  20. Quaternion applications for robot modeling

    Czech Academy of Sciences Publication Activity Database

    Ehrenberger, Zdeněk; Březina, Tomáš; Houška, P.

    Brno : VUT, 2002 - (Houfek, L.; Hlavoň, P.; Krejčí, P.), s. 1-8 ISBN 80-214-2109-6. [National conference with international participation Engineering Mechanics 2002. Svratka (CZ), 13.05.2002-16.05.2002] Institutional research plan: CEZ:AV0Z2076919 Keywords : robot * quaternion * modelling Subject RIV: JD - Computer Applications, Robotics

  1. A Gaussian Mixture MRF for Model-Based Iterative Reconstruction with Applications to Low-Dose X-ray CT

    CERN Document Server

    Zhang, Ruoqiao; Pal, Debashish; Thibault, Jean-Baptiste; Sauer, Ken D; Bouman, Charles A

    2016-01-01

    Markov random fields (MRFs) have been widely used as prior models in various inverse problems such as tomographic reconstruction. While MRFs provide a simple and often effective way to model the spatial dependencies in images, they suffer from the fact that parameter estimation is difficult. In practice, this means that MRFs typically have very simple structure that cannot completely capture the subtle characteristics of complex images. In this paper, we present a novel Gaussian mixture Markov random field model (GM-MRF) that can be used as a very expressive prior model for inverse problems such as denoising and reconstruction. The GM-MRF forms a global image model by merging together individual Gaussian-mixture models (GMMs) for image patches. In addition, we present a novel analytical framework for computing MAP estimates using the GM-MRF prior model through the construction of surrogate functions that result in a sequence of quadratic optimizations. We also introduce a simple but effective method to adjust...

  2. The Applicability of Two Strengths-Based Systemic Psychotherapy Models for Young People Following Type 1 Trauma

    Science.gov (United States)

    Coulter, Stephen

    2014-01-01

    This paper will consider the inter-relationship of a number of overlapping disciplinary theoretical concepts relevant to a strengths-based orientation, including well-being, salutogenesis, sense of coherence, quality of life and resilience. Psychological trauma will be referenced and the current evidence base for interventions with children and…

  3. Environmental Applications of Geochemical Modeling

    Science.gov (United States)

    Zhu, Chen; Anderson, Greg

    2002-05-01

    This book discusses the application of geochemical models to environmental practice and studies, through the use of numerous case studies of real-world environmental problems, such as acid mine drainage, pit lake chemistry, nuclear waste disposal, and landfill leachates. In each example the authors clearly define the environmental threat in question; explain how geochemical modeling may help solve the problem posed; and advise the reader how to prepare input files for geochemical modeling codes and interpret the results in terms of meeting regulatory requirements.

  4. P-Graph-based Workflow Modelling

    OpenAIRE

    József Tick

    2007-01-01

    Workflow modelling has been successfully introduced and implemented in severalapplication fields. Therefore, its significance has increased dramatically. Several work flowmodelling techniques have been published so far, out of which quite a number arewidespread applications. For instance the Petri-Net-based modelling has become popularpartly due to its graphical design and partly due to its correct mathematical background.The workflow modelling based on Unified Modelling Language is important...

  5. A Proposal of B to B Collaboration Process Model based on a Concept of Service and its Application to Energy Saving Service Business

    Science.gov (United States)

    Zhang, Qi; Kosaka, Michitaka; Shirahada, Kunio; Yabutani, Takashi

    This paper proposes a new framework for B to B collaboration process based on a concept of service. Service value, which gives users satisfaction for provided services, depends on the situation, user characteristics, and user objectives in seeking the service. Vargo proposed Service Dominant Logic (SDL), which determines service value according to “value in use”. This concept illustrates the importance of the relationship between the service itself and its situation. This relationship is analogous to electro-magnetic field theory in physics. We developed the concept of service fields to create service value based on an analogy of the electro-magnetic field. By applying this concept to B to B collaboration, a model of service value co-creation in the collaboration can be formulated. Then, the collaboration can be described by 4 steps of KIKI model (Knowledge sharing related to service system, Identification of service field, Knowledge creation for new service idea, Implementation of service idea). As its application to B to B collaboration, the energy saving service business is reported to demonstrate the validity of the proposed collaboration model. This concept can be applied to make a collaboration process effective.

  6. Physically-based 6-DoF Nodes Deformable Models: Application to Connective Tissues Simulation and Soft-Robots Control

    OpenAIRE

    Bosman, Julien

    2015-01-01

    The medical simulation is an increasingly active research field. Yet, despite the promising advance observed over the past years, the complete virtual patient’s model is yet to come. There are still many avenues for improvements, especially concerning the mechanical modeling of boundary conditions on anatomical structures.So far, most of the work has been dedicated to organs simulation, which are generally simulated alone. This raises a real problem as the role of the surrounding organs in th...

  7. Model-based segmentation

    OpenAIRE

    Heimann, Tobias; Delingette, Hervé

    2011-01-01

    This chapter starts with a brief introduction into model-based segmentation, explaining the basic concepts and different approaches. Subsequently, two segmentation approaches are presented in more detail: First, the method of deformable simplex meshes is described, explaining the special properties of the simplex mesh and the formulation of the internal forces. Common choices for image forces are presented, and how to evolve the mesh to adapt to certain structures. Second, the method of point...

  8. A Brief Introduction of the Achievements of Key Project Image-based Modeling and Rendering for Virtual Reality Applications

    Institute of Scientific and Technical Information of China (English)

    Jiaoying Shi; Zhanyi Hu; Enhua Wu; Qunsheng Peng

    2006-01-01

    @@ 1.Background The virtual reality (VR) technology is now at the frontier of modern information science.VR is based on computer graphics,computer vision,and other fresh air topics in today's computer technology.

  9. Quantitative phosphoproteomics using acetone-based peptide labeling: Method evaluation and application to a cardiac ischemia/reperfusion model

    OpenAIRE

    Wijeratne, Aruna B.; Manning, Janet R.; Schultz, Jo El J.; Greis, Kenneth D.

    2013-01-01

    Mass spectrometry (MS) techniques to globally profile protein phosphorylation in cellular systems that are relevant to physiological or pathological changes have been of significant interest in biological research. In this report, an MS-based strategy utilizing an inexpensive acetone-based peptide labeling technique known as reductive alkylation by acetone (RABA) for quantitative phosphoproteomics was explored to evaluate its capacity. Since the chemistry for RABA-labeling for phosphorylation...

  10. Automatic deployment of component-based applications

    OpenAIRE

    Lascu, Tudor Alexandru; Mauro, Jacopo; Zavattaro, Gianluigi

    2015-01-01

    International audience In distributed systems like those based on cloud or service-oriented frameworks, applications are typically assembled by deploying and connecting a large number of heterogeneous software components, spanning from fine-grained packages to coarse-grained complex services. Automation techniques and tools have been proposed to ease the deployment process of these complex system. By relying on a formal model of components, we describe a sound and complete algorithm for co...

  11. Yin-yang of space travel: lessons from the ground-based models of microgravity and their applications to disease and health for life on Earth

    Science.gov (United States)

    Kulkarni, A.; Yamauchi, K.; Hales, N.; Sundaresan, A.; Pellis, N.; Yamamoto, S.; Andrassy, R.

    Space flight environment has numerous clinical effects on human physiology; however, the advances made in physical and biological sciences have benefited humans on Earth. Space flight induces adverse effects on bone, muscle, cardiovascular, neurovestibular, gastrointestinal, and immune function. Similar pathophysiologic changes are also observed in aging with debilitating consequences. Anti-orthostatic tail-suspension (AOS) of rodents is an in vivo model to study many of these effects induced by the microgravity environment of space travel. Over the years AOS has been used by several researchers to study bone demineralization, muscle atrophy, neurovestibular and stress related effects. ecently we employed the AOS model in parallel with in vitro cell culture microgravity analog (Bioreactor) to document the decrease in immune function and its reversal by a nutritional countermeasure. We have modified the rodent model to study nutrient effects and benefits in a short period of time, usually within one to two weeks, in contrast to conventional aging research models which take several weeks to months to get the same results. This model has a potential for further development to study the role of nutrition in other pathophysiologies in an expedited manner. Using this model it is possible to evaluate the response of space travelers of various ages to microgravity stressors for long-term space travel. Hence this modified model will have significant impact on time and financial research budget. For the first time our group has documented a true potential immunonutritional countermeasure for the space flight induced effects on immune system (Clinical Nutrition 2002). Based on our nutritional and immunological studies we propose application of these microgravity analogs and its benefits and utility for nutritional effects on other physiologic parameters especially in aging. (Supported by NASA NCC8-168 grant, ADK)

  12. A finite-element-based perturbation model for the rotordynamic analysis of shrouded pump impellers: Part 1: Model development and applications

    Science.gov (United States)

    Baskharone, Erian A.

    1993-01-01

    This study concerns the rotor dynamic characteristics of fluid-encompassed rotors, with special emphasis on shrouded pump impellers. The core of the study is a versatile and categorically new finite-element-based perturbation model, which is based on a rigorous flow analysis and what we have generically termed the 'virtually' deformable finite-element approach. The model is first applied to the case of a smooth annular seal for verification purposes. The rotor excitation components, in this sample problem, give rise to a purely cylindrical, purely conical, and a simultaneous cylindrical/conical rotor whirl around the housing centerline. In all cases, the computed results are compared to existing experimental and analytical data involving the same seal geometry and operating conditions. Next, two labyrinth-seal configurations, which share the same tooth-to-tooth chamber geometry but differ in the total number of chambers, were investigated. The results, in this case, are compared to experimental measurements for both seal configurations. The focus is finally shifted to the shrouded-impeller problem, where the stability effects of the leakage flow in the shroud-to-housing secondary passage are investigated. To this end, the computational model is applied to a typical shrouded-impeller pump stage, fabricated and rotor dynamically tested by Sulzer Bros., and the results compared to those of a simplified 'bulk-flow' analysis and Sulzer Bros.' test data. In addition to assessing the computed rotor dynamic coefficients, the shrouded-impeller study also covers a controversial topic, namely that of the leakage-passage inlet swirl, which was previously cited as the origin of highly unconventional (resonance-like) trends of the fluid-exerted forces. In order to validate this claim, a 'microscopic' study of the fluid/shroud interaction mechanism is conducted, with the focus being on the structure of the perturbed flow field associated with the impeller whirl. The conclusions

  13. Wind-Climate Estimation Based on Mesoscale and Microscale Modeling: Statistical-Dynamical Downscaling for Wind Energy Applications

    DEFF Research Database (Denmark)

    Badger, Jake; Frank, Helmut; Hahmann, Andrea N.;

    2014-01-01

    This paper demonstrates that a statistical dynamical method can be used to accurately estimate the wind climate at a wind farm site. In particular, postprocessing of mesoscale model output allows an efficient calculation of the local wind climate required for wind resource estimation at a wind tu...

  14. Application of a dynamic population-based model for evaluation of exposure reduction strategies in the baking industry

    Energy Technology Data Exchange (ETDEWEB)

    Meijster, Tim; Tielemans, Erik [TNO Quality of Life, Business unit Quality and Safety, Zeist (Netherlands); Warren, Nick [Health and Safety Laboratory, Harpur Hill, Buxton, Derbyshire (United Kingdom); Heederik, Dick, E-mail: Tim.meijster@tno.n [Utrecht University, Institute of Risk Assessment Sciences, Division of Environmental Epidemiology, Utrecht (Netherlands)

    2009-02-01

    Recently a dynamic population model was developed that simulates a population of bakery workers longitudinally through time and tracks the development of work-related sensitisation and respiratory symptoms in each worker. Input for this model comes from cross-sectional and longitudinal epidemiological studies which allowed estimation of exposure response relationships and disease transition probabilities This model allows us to study the development of diseases and transitions between disease states over time in relation to determinants of disease including flour dust and/or allergen exposure. Furthermore it enables more realistic modelling of the health impact of different intervention strategies at the workplace (e.g. changes in exposure may take several years to impact on ill-health and often occur as a gradual trend). A large dataset of individual full-shift exposure measurements and real-time exposure measurements were used to obtain detailed insight into the effectiveness of control measures and other determinants of exposure. Given this information a population wide reduction of the median exposure with 50% was evaluated in this paper.

  15. Application of a dynamic population-based model for evaluation of exposure reduction strategies in the baking industry

    International Nuclear Information System (INIS)

    Recently a dynamic population model was developed that simulates a population of bakery workers longitudinally through time and tracks the development of work-related sensitisation and respiratory symptoms in each worker. Input for this model comes from cross-sectional and longitudinal epidemiological studies which allowed estimation of exposure response relationships and disease transition probabilities This model allows us to study the development of diseases and transitions between disease states over time in relation to determinants of disease including flour dust and/or allergen exposure. Furthermore it enables more realistic modelling of the health impact of different intervention strategies at the workplace (e.g. changes in exposure may take several years to impact on ill-health and often occur as a gradual trend). A large dataset of individual full-shift exposure measurements and real-time exposure measurements were used to obtain detailed insight into the effectiveness of control measures and other determinants of exposure. Given this information a population wide reduction of the median exposure with 50% was evaluated in this paper.

  16. Geophysical Model Applications for Monitoring

    Energy Technology Data Exchange (ETDEWEB)

    Pasyanos, M; Walter, W; Tkalcic, H; Franz, G; Gok, R; Rodgers, A

    2005-07-11

    Geophysical models constitute an important component of calibration for nuclear explosion monitoring. We will focus on four major topics and their applications: (1) surface wave models, (2) receiver function profiles, (3) regional tomography models, and (4) stochastic geophysical models. First, we continue to improve upon our surface wave model by adding more paths. This has allowed us to expand the region to all of Eurasia and into Africa, increase the resolution of our model, and extend results to even shorter periods (7 sec). High-resolution models exist for the Middle East and the YSKP region. The surface wave results can be inverted either alone, or in conjunction with other data, to derive models of the crust and upper mantle structure. One application of the group velocities is to construct phase-matched filters in combination with regional surface-wave magnitude formulas to improve the mb:Ms discriminant and extend it to smaller magnitude events. Next, we are using receiver functions, in joint inversions with the surface waves, to produce profiles directly under seismic stations throughout the region. In the past year, we have been focusing on deployments throughout the Middle East, including the Arabian Peninsula and Turkey. By assembling the results from many stations, we can see how regional seismic phases are affected by complicated upper mantle structure, including lithospheric thickness and anisotropy. The next geophysical model item, regional tomography models, can be used to predict regional travel times such as Pn and Sn. The times derived by the models can be used as a background model for empirical measurements or, where these don't exist, simply used as is. Finally, we have been exploring methodologies such as Markov Chain Monte Carlo (MCMC) to generate data-driven stochastic models. We have applied this technique to the YSKP region using surface wave dispersion data, body wave travel time data, receiver functions, and gravity data. The

  17. Security Assessment of Web Based Distributed Applications

    Directory of Open Access Journals (Sweden)

    Catalin BOJA

    2010-01-01

    Full Text Available This paper presents an overview about the evaluation of risks and vulnerabilities in a web based distributed application by emphasizing aspects concerning the process of security assessment with regards to the audit field. In the audit process, an important activity is dedicated to the measurement of the characteristics taken into consideration for evaluation. From this point of view, the quality of the audit process depends on the quality of assessment methods and techniques. By doing a review of the fields involved in the research process, the approach wants to reflect the main concerns that address the web based distributed applications using exploratory research techniques. The results show that many are the aspects which must carefully be worked with, across a distributed system and they can be revealed by doing a depth introspective analyze upon the information flow and internal processes that are part of the system. This paper reveals the limitations of a non-existing unified security risk assessment model that could prevent such risks and vulnerabilities debated. Based on such standardize models, secure web based distributed applications can be easily audited and many vulnerabilities which can appear due to the lack of access to information can be avoided.

  18. A thermodynamically-based model for predicting microbial growth and community composition coupled to system geochemistry: Application to uranium bioreduction

    Science.gov (United States)

    Istok, J. D.; Park, M.; Michalsen, M.; Spain, A. M.; Krumholz, L. R.; Liu, C.; McKinley, J.; Long, P.; Roden, E.; Peacock, A. D.; Baldwin, B.

    2010-03-01

    'Bioimmobilization' of redox-sensitive heavy metals and radionuclides is being investigated as a way to remediate contaminated groundwater and sediments. In one approach, growth-limiting substrates are added to the subsurface to stimulate the activity of targeted groups of indigenous microorganisms and create conditions favorable for the microbially-mediated reductive precipitation ('bioreduction') of targeted contaminants. We present a theoretical framework for modeling this process that modifies conventional geochemical reaction path modeling to include thermodynamic descriptions for microbial growth and may be called biogeochemical reaction path modeling. In this approach, the actual microbial community is represented by a synthetic microbial community consisting of a collection of microbial groups; each with a unique growth equation that couples a specific pair of energy yielding redox reactions. The growth equations and their computed standard-state free energy yields are appended to the thermodynamic database used in conventional geochemical reaction path modeling, providing a direct coupling between chemical species participating in both microbial growth and geochemical reactions. To compute the biogeochemical reaction paths, growth substrates are reacted incrementally with the defined geochemical environment and the coupled equations are solved simultaneously to predict reaction paths that display changing microbial biomass, community composition (i.e. the fraction of total biomass in each microbial group), and the aqueous and mineral composition of the system, including aqueous speciation and oxidation state of the targeted contaminants. The approach, with growth equations derived from the literature using well-known bioenergetics principles, was used to predict the results of a laboratory microcosm experiment and an in situ field experiment that investigated the bioreduction of uranium. Predicted effects of ethanol or acetate addition on uranium

  19. MIRAGE: a functional genomics-based approach for metabolic network model reconstruction and its application to cyanobacteria networks

    OpenAIRE

    Vitkin, Edward; Shlomi, Tomer

    2012-01-01

    Genome-scale metabolic network reconstructions are considered a key step in quantifying the genotype-phenotype relationship. We present a novel gap-filling approach, MetabolIc Reconstruction via functionAl GEnomics (MIRAGE), which identifies missing network reactions by integrating metabolic flux analysis and functional genomics data. MIRAGE's performance is demonstrated on the reconstruction of metabolic network models of E. coli and Synechocystis sp. and validated via existing networks for ...

  20. Parametric estimation of covariance function in Gaussian-process based Kriging models. Application to uncertainty quantification for computer experiments

    OpenAIRE

    Bachoc, François

    2013-01-01

    The parametric estimation of the covariance function of a Gaussian process is studied, in the framework of the Kriging model. Maximum Likelihood and Cross Validation estimators are considered. The correctly specified case, in which the covariance function of the Gaussian process does belong to the parametric set used for estimation, is first studied in an increasing-domain asymptotic framework. The sampling considered is a randomly perturbed multidimensional regular grid. Consistency and asym...

  1. Estimation of the physico-chemical parameters of materials based on rare earth elements with the application of computational model

    Science.gov (United States)

    Mamaev, K.; Obkhodsky, A.; Popov, A.

    2016-01-01

    Computational model, technique and the basic principles of operation program complex for quantum-chemical calculations of material's physico-chemical parameters with rare earth elements are discussed. The calculating system is scalable and includes CPU and GPU computational resources. Control and operation of computational jobs and also Globus Toolkit 5 software provides the possibility to join computer users in a unified system of data processing with peer-to-peer architecture. CUDA software is used to integrate graphic processors into calculation system.

  2. Early FDI Based on Residuals Design According to the Analysis of Models of Faults: Application to DAMADICS

    OpenAIRE

    Noureddine Guersi; Dimitri Lefebvre; Yahia Kourd

    2011-01-01

    The increased complexity of plants and the development of sophisticated control systems have encouraged the parallel development of efficient rapid fault detection and isolation (FDI) systems. FDI in industrial system has lately become of great significance. This paper proposes a new technique for short time fault detection and diagnosis in nonlinear dynamic systems with multi inputs and multi outputs. The main contribution of this paper is to develop a FDI schema according to reference model...

  3. Application of GSO for Load Allocation between Hydropower Units and Its Model Analysis based on Multi-objective

    OpenAIRE

    Liying Wang; Linming Zhao; Hongyan Yan

    2012-01-01

    The optimum load distribution between the hydropower units is an effective measure for reducing the total water rate and increasing the energy output of a hydropower station, and it is becoming a more interesting studying topic. In order to increase the economic benefit of the station, a multi-objective optimization model of load allocation between hydropower units is established in accordance with the characteristics and particularity of the hydropower station, and the minimum water rate of ...

  4. A Critical Review on Wind Turbine Power Curve Modelling Techniques and Their Applications in Wind Based Energy Systems

    OpenAIRE

    Sohoni, Vaishali; Gupta, S. C.; R. K. Nema

    2016-01-01

    Power curve of a wind turbine depicts the relationship between output power and hub height wind speed and is an important characteristic of the turbine. Power curve aids in energy assessment, warranty formulations, and performance monitoring of the turbines. With the growth of wind industry, turbines are being installed in diverse climatic conditions, onshore and offshore, and in complex terrains causing significant departure of these curves from the warranted values. Accurate models of power...

  5. Extraction of Desired Signal Based on AR Model with Its Application to Atrial Activity Estimation in Atrial Fibrillation

    Science.gov (United States)

    Wang, Gang; Rao, Ni-ni; Shepherd, Simon J.; Beggs, Clive B.

    2008-12-01

    The use of electrocardiograms (ECGs) to diagnose and analyse atrial fibrillation (AF) has received much attention recently. When studying AF, it is important to isolate the atrial activity (AA) component of the ECG plot. We present a new autoregressive (AR) model for semiblind source extraction of the AA signal. Previous researchers showed that one could extract a signal with the smallest normalized mean square prediction error (MSPE) as the first output from linear mixtures by minimizing the MSPE. However the extracted signal will be not always the desired one even if the AR model parameters of one source signal are known. We introduce a new cost function, which caters for the specific AR model parameters, to extract the desired source. Through theoretical analysis and simulation we demonstrate that this algorithm can extract any desired signal from mixtures provided that its AR parameters are first obtained. We use this approach to extract the AA signal from 12-lead surface ECG signals for hearts undergoing AF. In our methodology we roughly estimated the AR parameters from the fibrillatory wave segment in the V1 lead, and then used this algorithm to extract the AA signal. We validate our approach using real-world ECG data.

  6. Study and Application of Reinforcement Learning in Cooperative Strategy of the Robot Soccer Based on BDI Model

    Directory of Open Access Journals (Sweden)

    Wu Bo-ying

    2009-11-01

    Full Text Available The dynamic cooperation model of multi-Agent is formed by combining reinforcement learning with BDI model. In this model, the concept of the individual optimization loses its meaning, because the repayment of each Agent dose not only depend on itsself but also on the choice of other Agents. All Agents can pursue a common optimum solution and try to realize the united intention as a whole to a maximum limit. The robot moves to its goal, depending on the present positions of the other robots that cooperate with it and the present position of the ball. One of these robots cooperating with it is controlled to move by man with a joystick. In this way, Agent can be ensured to search for each state-action as frequently as possible when it carries on choosing movements, so as to shorten the time of searching for the movement space so that the convergence speed of reinforcement learning can be improved. The validity of the proposed cooperative strategy for the robot soccer has been proved by combining theoretical analysis with simulation robot soccer match (11vs11 .

  7. Applications of a thermal-based two-source energy balance model using Priestley-Taylor approach for surface temperature partitioning under advective conditions

    Science.gov (United States)

    Song, Lisheng; Kustas, William P.; Liu, Shaomin; Colaizzi, Paul D.; Nieto, Hector; Xu, Ziwei; Ma, Yanfei; Li, Mingsong; Xu, Tongren; Agam, Nurit; Tolk, Judy A.; Evett, Steven R.

    2016-09-01

    In this study ground measured soil and vegetation component temperatures and composite temperature from a high spatial resolution thermal camera and a network of thermal-IR sensors collected in an irrigated maize field and in an irrigated cotton field are used to assess and refine the component temperature partitioning approach in the Two-Source Energy Balance (TSEB) model. A refinement to TSEB using a non-iterative approach based on the application of the Priestley-Taylor formulation for surface temperature partitioning and estimating soil evaporation from soil moisture observations under advective conditions (TSEB-A) was developed. This modified TSEB formulation improved the agreement between observed and modeled soil and vegetation temperatures. In addition, the TSEB-A model output of evapotranspiration (ET) and the components evaporation (E), transpiration (T) when compared to ground observations using the stable isotopic method and eddy covariance (EC) technique from the HiWATER experiment and with microlysimeters and a large monolithic weighing lysimeter from the BEAREX08 experiment showed good agreement. Difference between the modeled and measured ET measurements were less than 10% and 20% on a daytime basis for HiWATER and BEAREX08 data sets, respectively. The TSEB-A model was found to accurately reproduce the temporal dynamics of E, T and ET over a full growing season under the advective conditions existing for these irrigated crops located in arid/semi-arid climates. With satellite data this TSEB-A modeling framework could potentially be used as a tool for improving water use efficiency and conservation practices in water limited regions. However, TSEB-A requires soil moisture information which is not currently available routinely from satellite at the field scale.

  8. Multi-scale modeling of the thermo-hydro- mechanical behaviour of heterogeneous materials. Application to cement-based materials under severe loads

    International Nuclear Information System (INIS)

    The work of modeling presented here relates to the study of the thermo-hydro- mechanical behaviour of porous materials based on hydraulic binder such as concrete, High Performance Concrete or more generally cement-based materials. This work is based on the exploitation of the Digital Concrete model, of the finite element code Symphonie developed in the Scientific and Technical Centre for Building (CSTB), in coupling with the homogenization methods to obtain macroscopic behaviour laws drawn from the Micro-Macro relations. Scales of investigation, macroscopic and microscopic, has been exploited by simulation in order to allow the comprehension fine of the behaviour of cement-based materials according to thermal, hydrous and mechanical loads. It appears necessary to take into account various scales of modeling. In order to study the behaviour of the structure, we are brought to reduce the scale of investigation to study the material more particularly. The research tasks presented suggest a new approach for the identification of the multi-physic behaviour of materials by simulation. In complement of the purely experimental approach, based on observations on the sample with measurements of the apparent parameters on the macroscopic scale, this new approach allows to obtain the fine analysis of elementary mechanisms in acting within the material. These elementary mechanisms are at the origin of the evolution of the macroscopic parameters measured in experimental tests. In this work, coefficients of the thermo-hydro-mechanical behaviour law of porous materials and the equivalent hydraulic conductivity were obtained by a multi-scales approach. Applications has been carried out on the study of the damaged behaviour of cement-based materials, in the objective to determine the elasticity tensor and the permeability tensor of a High Performance Concrete at high temperatures under a mechanical load. Also, the study of the strain evolution of cement-based materials at low

  9. Design of a multi-model observer-based estimator for Fault Detection and Isolation (FDI strategy: application to a chemical reactor

    Directory of Open Access Journals (Sweden)

    Y. Chetouani

    2008-12-01

    Full Text Available This study presents a FDI strategy for nonlinear dynamic systems. It shows a methodology of tackling the fault detection and isolation issue by combining a technique based on the residuals signal and a technique using the multiple Kalman filters. The usefulness of this combination is the on-line implementation of the set of models, which represents the normal mode and all dynamics of faults, if the statistical decision threshold on the residuals exceeds a fixed value. In other cases, one Extended Kalman Filter (EKF is enough to estimate the process state. After describing the system architecture and the proposed FDI methodology, we present a realistic application in order to show the technique's potential. An algorithm is described and applied to a chemical process like a perfectly stirred chemical reactor functioning in a semi-batch mode. The chemical reaction used is an oxido reduction one, the oxidation of sodium thiosulfate by hydrogen peroxide.

  10. Value co-creation model based on IS application capabilities%基于 IS应用能力的价值共同创造模型

    Institute of Scientific and Technical Information of China (English)

    朱树婷; 仲伟俊; 梅姝娥

    2014-01-01

    为研究电子商务背景下企业和顾客之间的价值创造,从关系观的理论视角构建基于IS应用能力的价值共同创造模型,分析IS应用能力、关系资产、顾客敏捷性和关系价值的构念和维度,应用实证方法验证理论假设。实证研究验证了顾客敏捷性在价值创造的中介作用、IS应用能力对关系资产的互补作用及其对关系价值产生的影响。基于IS应用能力的价值共同创造模型拓展了服务主导逻辑下的价值创造框架,新的理论模型和实证结论解释了电子商务环境下企业通过IS应用能力与顾客进行价值共同创造的机理,为进一步的理论发展和企业实践提供指导。%To illustrate how firms and customers co-create value in business to business B2B e-commerce an integrated value co-creation model is proposed based on information systems IS application capabilities from the relational view.IS application capabilities relational assets customer agility and relational value are constructed and tested by empirical analysis.The empirical research tests and verifies the mediating effect of customer agility and the interactions of IS application capabilities and relational assets as well as their effect on relational value. This model expands the research framework of value co-creation in service dominant logic and reveals the mechanism of how firms and customers co-create value in B2B e-commerce based on IS application capabilities which provides the basis for further theory development and a practice guide.

  11. Voxel-Based LIDAR Analysis and Applications

    Science.gov (United States)

    Hagstrom, Shea T.

    One of the greatest recent changes in the field of remote sensing is the addition of high-quality Light Detection and Ranging (LIDAR) instruments. In particular, the past few decades have been greatly beneficial to these systems because of increases in data collection speed and accuracy, as well as a reduction in the costs of components. These improvements allow modern airborne instruments to resolve sub-meter details, making them ideal for a wide variety of applications. Because LIDAR uses active illumination to capture 3D information, its output is fundamentally different from other modalities. Despite this difference, LIDAR datasets are often processed using methods appropriate for 2D images and that do not take advantage of its primary virtue of 3-dimensional data. It is this problem we explore by using volumetric voxel modeling. Voxel-based analysis has been used in many applications, especially medical imaging, but rarely in traditional remote sensing. In part this is because the memory requirements are substantial when handling large areas, but with modern computing and storage this is no longer a significant impediment. Our reason for using voxels to model scenes from LIDAR data is that there are several advantages over standard triangle-based models, including better handling of overlapping surfaces and complex shapes. We show how incorporating system position information from early in the LIDAR point cloud generation process allows radiometrically-correct transmission and other novel voxel properties to be recovered. This voxelization technique is validated on simulated data using the Digital Imaging and Remote Sensing Image Generation (DIRSIG) software, a first-principles based ray-tracer developed at the Rochester Institute of Technology. Voxel-based modeling of LIDAR can be useful on its own, but we believe its primary advantage is when applied to problems where simpler surface-based 3D models conflict with the requirement of realistic geometry. To

  12. Quantitative phosphoproteomics using acetone-based peptide labeling: method evaluation and application to a cardiac ischemia/reperfusion model.

    Science.gov (United States)

    Wijeratne, Aruna B; Manning, Janet R; Schultz, Jo El J; Greis, Kenneth D

    2013-10-01

    Mass spectrometry (MS) techniques to globally profile protein phosphorylation in cellular systems that are relevant to physiological or pathological changes have been of significant interest in biological research. An MS-based strategy utilizing an inexpensive acetone-based peptide-labeling technique known as reductive alkylation by acetone (RABA) for quantitative phosphoproteomics was explored to evaluate its capacity. Because the chemistry for RABA labeling for phosphorylation profiling had not been previously reported, it was first validated using a standard phosphoprotein and identical phosphoproteomes from cardiac tissue extracts. A workflow was then utilized to compare cardiac tissue phosphoproteomes from mouse hearts not expressing FGF2 versus hearts expressing low-molecular-weight fibroblast growth factor-2 (LMW FGF2) to relate low-molecular-weight fibroblast growth factor-2 (LMW FGF2)-mediated cardioprotective phenomena induced by ischemia/reperfusion injury of hearts, with downstream phosphorylation changes in LMW FGF2 signaling cascades. Statistically significant phosphorylation changes were identified at 14 different sites on 10 distinct proteins, including some with mechanisms already established for LMW FGF2-mediated cardioprotective signaling (e.g., connexin-43), some with new details linking LMW FGF2 to the cardioprotective mechanisms (e.g., cardiac myosin binding protein C or cMyBPC), and also several new downstream effectors not previously recognized for cardio-protective signaling by LMW FGF2. Additionally, one of the phosphopeptides, cMyBPC/pSer-282, identified was further verified with site-specific quantification using an SRM (selected reaction monitoring)-based approach that also relies on isotope labeling of a synthetic phosphopeptide with deuterated acetone as an internal standard. Overall, this study confirms that the inexpensive acetone-based peptide labeling can be used in both exploratory and targeted quantification phosphoproteomic

  13. Application of flood risk modelling in a web-based geospatial decision support tool for coastal adaptation to climate change

    OpenAIRE

    P. J. Knight; T. Prime; J M Brown; Morrissey, K; Plater, A.J.

    2015-01-01

    A pressing problem facing coastal decision makers is the conversion of "high-level" but plausible climate change assessments into an effective basis for climate change adaptation at the local scale. Here, we describe a web-based, geospatial decision support tool (DST) that provides an assessment of the potential flood risk for populated coastal lowlands arising from future sea-level rise, coastal storms, and high river flows. This DST has been developed to support operationa...

  14. Application of flood risk modelling in a web-based geospatial decision support tool for coastal adaptation to climate change

    OpenAIRE

    P. J. Knight; T. Prime; J M Brown; Morrissey, K; Plater, A.J.

    2015-01-01

    A pressing problem facing coastal decision makers is the conversion of "high level" but plausible climate change assessments into an effective basis for climate change adaptation at the local scale. Here, we describe a web-based, geospatial decision-support tool (DST) that provides an assessment of the potential flood risk for populated coastal lowlands arising from future sea-level rise, coastal storms and high river flows. This DST has been d...

  15. Application of a Model to Evaluate Infrared Exposure Limits in Aluminum Foundries Based on Threshold Temperature in the Range of 770-1400 nm

    Directory of Open Access Journals (Sweden)

    FARAMARZ MADJIDI

    2015-10-01

    Full Text Available High intensity optical radiation can cause damage to the eye and intense radiation in the range of 770-1400 nm can cause thermal retinal damage. In the workplaces where there are high temperature sources, the workers in front of these hot  sources without bright light maybe exposed to the  intense IR radiation, thus regular measurement of these radiations seems crucial. Measurement of IR radiations by radiometer in  specific wavelength ranges is elusive. Moreover, when radiometers are used, the correct application of the recommended exposure limits requires knowledge of spectral radiance which seems sophisticated for hygienists. The main objective of the present study is applying a model to express retinal thermal injury in terms of temperature for molten aluminum ovens in an aluminum foundry that emit optical radiation without visible light. In the proposed model, ACGIH TLVs for retinal thermal injury in the range of 770 to 1400 nm was used where source luminance was under 0.01 cd/cm2. Also, by using the output results of this proposed model it is possible to present a new chart for evaluation of exposure to IR for hot sources based on Threshold Temperature.

  16. Lévy-based growth models

    DEFF Research Database (Denmark)

    Jónsdóttir, Kristjana Ýr; Schmiegel, Jürgen; Jensen, Eva Bjørn Vedel

    2008-01-01

    In the present paper, we give a condensed review, for the nonspecialist reader, of a new modelling framework for spatio-temporal processes, based on Lévy theory. We show the potential of the approach in stochastic geometry and spatial statistics by studying Lévy-based growth modelling of planar...... objects. The growth models considered are spatio-temporal stochastic processes on the circle. As a by product, flexible new models for space–time covariance functions on the circle are provided. An application of the Lévy-based growth models to tumour growth is discussed....

  17. Some applications of neural networks in microwave modeling

    Directory of Open Access Journals (Sweden)

    Milovanović Bratislav D.

    2003-01-01

    Full Text Available This paper presents some applications of neural networks in the microwave modeling. The applications are related to modeling of either passive or active structures and devices. Modeling is performed using not only simple multilayer perception network (MLP but also advanced knowledge based neural network (KBNN structures.

  18. Optimal multi-agent path planning for fast inverse modeling in UAV-based flood sensing applications

    KAUST Repository

    Abdelkader, Mohamed

    2014-05-01

    Floods are the most common natural disasters, causing thousands of casualties every year in the world. In particular, flash flood events are particularly deadly because of the short timescales on which they occur. Unmanned air vehicles equipped with mobile microsensors could be capable of sensing flash floods in real time, saving lives and greatly improving the efficiency of the emergency response. However, of the main issues arising with sensing floods is the difficulty of planning the path of the sensing agents in advance so as to obtain meaningful data as fast as possible. In this particle, we present a fast numerical scheme to quickly compute the trajectories of a set of UAVs in order to maximize the accuracy of model parameter estimation over a time horizon. Simulation results are presented, a preliminary testbed is briefly described, and future research directions and problems are discussed. © 2014 IEEE.

  19. INCLUSION RATIO BASED ESTIMATOR FOR THE MEAN LENGTH OF THE BOOLEAN LINE SEGMENT MODEL WITH AN APPLICATION TO NANOCRYSTALLINE CELLULOSE

    Directory of Open Access Journals (Sweden)

    Mikko Niilo-Rämä

    2014-06-01

    Full Text Available A novel estimator for estimating the mean length of fibres is proposed for censored data observed in square shaped windows. Instead of observing the fibre lengths, we observe the ratio between the intensity estimates of minus-sampling and plus-sampling. It is well-known that both intensity estimators are biased. In the current work, we derive the ratio of these biases as a function of the mean length assuming a Boolean line segment model with exponentially distributed lengths and uniformly distributed directions. Having the observed ratio of the intensity estimators, the inverse of the derived function is suggested as a new estimator for the mean length. For this estimator, an approximation of its variance is derived. The accuracies of the approximations are evaluated by means of simulation experiments. The novel method is compared to other methods and applied to real-world industrial data from nanocellulose crystalline.

  20. Pattern-based Automatic Translation of Structured Power System Data to Functional Models for Decision Support Applications

    DEFF Research Database (Denmark)

    Heussen, Kai; Weckesser, Johannes Tilman Gabriel; Kullmann, Daniel

    2013-01-01

    Improved information and insight for decision support in operations and design are central promises of a smart grid. Well-structured information about the composition of power systems is increasingly becoming available in the domain, e.g. due to standard information models (e.g. CIM or IEC61850) or...... otherwise structured databases. More measurements and data do not automatically improve decisions, but there is an opportunity to capitalize on this information for decision support. With suitable reasoning strategies data can be contextualized and decision-relevant events can be promoted and identified....... This paper presents an approach to link available structured power system data directly to a functional representation suitable for diagnostic reasoning. The translation method is applied to test cases also illustrating decision support....

  1. Application of the Pareto Principle in Rapid Application Development Model

    Directory of Open Access Journals (Sweden)

    Vishal Pandey

    2013-06-01

    Full Text Available the Pareto principle or most popularly termed as the 80/20 rule is one of the well-known theories in the field of economics. This rule of thumb was named after the great economist Vilferdo Pareto. The Pareto principle was proposed by a renowned management consultant Joseph M Juran. The rule states that 80% of the required work can be completed in 20% of the time allotted. The idea is to apply this rule of thumb in the Rapid Application Development (RAD Process model of software engineering. The Rapid application development model integrates end-user in the development using iterative prototyping emphasizing on delivering a series of fully functional prototype to designated user experts. During the application of Pareto Principle the other concepts like the Pareto indifference curve and Pareto efficiency also come into the picture. This enables the development team to invest major amount of time focusing on the major functionalities of the project as per the requirement prioritizationof the customer. The paper involves an extensive study on different unsatisfactory projects in terms of time and financial resources and the reasons of failures are analyzed. Based on the possible reasons offailure, a customized RAD model is proposed integrating the 80/20 rule and advanced software development strategies to develop and deploy excellent quality software product in minimum time duration. The proposed methodology is such that its application will directly affect the quality of the end product for the better.

  2. Application of Multicast-based Video Conference on CERNET Backbone

    Institute of Scientific and Technical Information of China (English)

    2002-01-01

    Multicast-based video conference is a representative application in advanced network. In multi-point video conference using multicast can get better efficiency facilitated by inner-group broadcast mechanism. In the application, the multicast-based network resources assignment, management and security should be considered together. This paper presents a framework model of multicast-based video conferencing application with three layers. And a practical multicast-based video conferencing is implemented in CERNET(China Education and Research Network) backbone. The practice is valuable for the development of multicast-based video conferencing application in China.

  3. The application of GIS based decision-tree models for generating the spatial distribution of hydromorphic organic landscapes in relation to digital terrain data

    Directory of Open Access Journals (Sweden)

    R. Bou Kheir

    2010-06-01

    Full Text Available Accurate information about organic/mineral soil occurrence is a prerequisite for many land resources management applications (including climate change mitigation. This paper aims at investigating the potential of using geomorphometrical analysis and decision tree modeling to predict the geographic distribution of hydromorphic organic landscapes in unsampled area in Denmark. Nine primary (elevation, slope angle, slope aspect, plan curvature, profile curvature, tangent curvature, flow direction, flow accumulation, and specific catchment area and one secondary (steady-state topographic wetness index topographic parameters were generated from Digital Elevation Models (DEMs acquired using airborne LIDAR (Light Detection and Ranging systems. They were used along with existing digital data collected from other sources (soil type, geological substrate and landscape type to explain organic/mineral field measurements in hydromorphic landscapes of the Danish area chosen. A large number of tree-based classification models (186 were developed using (1 all of the parameters, (2 the primary DEM-derived topographic (morphological/hydrological parameters only, (3 selected pairs of parameters and (4 excluding each parameter one at a time from the potential pool of predictor parameters. The best classification tree model (with the lowest misclassification error and the smallest number of terminal nodes and predictor parameters combined the steady-state topographic wetness index and soil type, and explained 68% of the variability in organic/mineral field measurements. The overall accuracy of the predictive organic/inorganic landscapes' map produced (at 1:50 000 cartographic scale using the best tree was estimated to be ca. 75%. The proposed classification-tree model is relatively simple, quick, realistic and practical, and it can be applied to other areas, thereby providing a tool to facilitate the implementation of pedological/hydrological plans for conservation

  4. The application of GIS based decision-tree models for generating the spatial distribution of hydromorphic organic landscapes in relation to digital terrain data

    Directory of Open Access Journals (Sweden)

    R. Bou Kheir

    2010-01-01

    Full Text Available Accurate information about soil organic carbon (SOC, presented in a spatially form, is prerequisite for many land resources management applications (including climate change mitigation. This paper aims to investigate the potential of using geomorphometrical analysis and decision tree modeling to predict the geographic distribution of hydromorphic organic landscapes at unsampled area in Denmark. Nine primary (elevation, slope angle, slope aspect, plan curvature, profile curvature, tangent curvature, flow direction, flow accumulation, and specific catchment area and one secondary (steady-state topographic wetness index topographic parameters were generated from Digital Elevation Models (DEMs acquired using airborne LIDAR (Light Detection and Ranging systems. They were used along with existing digital data collected from other sources (soil type, geological substrate and landscape type to statistically explain SOC field measurements in hydromorphic landscapes of the chosen Danish area. A large number of tree-based classification models (186 were developed using (1 all of the parameters, (2 the primary DEM-derived topographic (morphological/hydrological parameters only, (3 selected pairs of parameters and (4 excluding each parameter one at a time from the potential pool of predictor parameters. The best classification tree model (with the lowest misclassification error and the smallest number of terminal nodes and predictor parameters combined the steady-state topographic wetness index and soil type, and explained 68% of the variability in field SOC measurements. The overall accuracy of the produced predictive SOC map (at 1:50 000 cartographic scale using the best tree was estimated to be ca. 75%. The proposed classification-tree model is relatively simple, quick, realistic and practical, and it can be applied to other areas, thereby providing a tool to help with the implementation of pedological/hydrological plans for conservation and sustainable

  5. A robust hybrid model integrating enhanced inputs based extreme learning machine with PLSR (PLSR-EIELM) and its application to intelligent measurement.

    Science.gov (United States)

    He, Yan-Lin; Geng, Zhi-Qiang; Xu, Yuan; Zhu, Qun-Xiong

    2015-09-01

    In this paper, a robust hybrid model integrating an enhanced inputs based extreme learning machine with the partial least square regression (PLSR-EIELM) was proposed. The proposed PLSR-EIELM model can overcome two main flaws in the extreme learning machine (ELM), i.e. the intractable problem in determining the optimal number of the hidden layer neurons and the over-fitting phenomenon. First, a traditional extreme learning machine (ELM) is selected. Second, a method of randomly assigning is applied to the weights between the input layer and the hidden layer, and then the nonlinear transformation for independent variables can be obtained from the output of the hidden layer neurons. Especially, the original input variables are regarded as enhanced inputs; then the enhanced inputs and the nonlinear transformed variables are tied together as the whole independent variables. In this way, the PLSR can be carried out to identify the PLS components not only from the nonlinear transformed variables but also from the original input variables, which can remove the correlation among the whole independent variables and the expected outputs. Finally, the optimal relationship model of the whole independent variables with the expected outputs can be achieved by using PLSR. Thus, the PLSR-EIELM model is developed. Then the PLSR-EIELM model served as an intelligent measurement tool for the key variables of the Purified Terephthalic Acid (PTA) process and the High Density Polyethylene (HDPE) process. The experimental results show that the predictive accuracy of PLSR-EIELM is stable, which indicate that PLSR-EIELM has good robust character. Moreover, compared with ELM, PLSR, hierarchical ELM (HELM), and PLSR-ELM, PLSR-EIELM can achieve much smaller predicted relative errors in these two applications. PMID:26112928

  6. Adjoint sensitivity analysis of dynamic reliability models based on Markov chains - II: Application to IFMIF reliability assessment

    International Nuclear Information System (INIS)

    In Part II of this work, the adjoint sensitivity analysis procedure developed in Part I is applied to perform sensitivity analysis of several dynamic reliability models of systems of increasing complexity, culminating with the consideration of the International Fusion Materials Irradiation Facility (IFMIF) accelerator system. Section II presents the main steps of a procedure for the automated generation of Markov chains for reliability analysis, including the abstraction of the physical system, construction of the Markov chain, and the generation and solution of the ensuing set of differential equations; all of these steps have been implemented in a stand-alone computer code system called QUEFT/MARKOMAG-S/MCADJSEN. This code system has been applied to sensitivity analysis of dynamic reliability measures for a paradigm '2-out-of-3' system comprising five components and also to a comprehensive dynamic reliability analysis of the IFMIF accelerator system facilities for the average availability and, respectively, the system's availability at the final mission time. The QUEFT/MARKOMAG-S/MCADJSEN has been used to efficiently compute sensitivities to 186 failure and repair rates characterizing components and subsystems of the first-level fault tree of the IFMIF accelerator system. (authors)

  7. Application of the Western-based adjuvant online model to Korean colon cancer patients; a single institution experience

    International Nuclear Information System (INIS)

    Adjuvant Online (AOL) is web-accessible risk-assessment model that predicts the mortality and the benefits of adjuvant therapy. AOL has never been validated for Asian colon cancer patients. Using the Yonsei Tumor Registry database, patients who were treated within the Yonsei University Health System between 1990 and 2005 for T1-4, N0-2, and M0 colon cancer were included in the calculations for survival. Observed and predicted 5-year overall survival was compared for each patient. The median age of the study population of 1431 patients was 60 years (range, 15–87 years), and the median follow-up duration was 7.9 years (range, 0.06–19.8 years). The predicted 5-year overall survival rate (77.7%) and observed survival (79.5%) was not statistically different (95% Confidential interval, 76.3–81.5) in all patients. Predicted outcomes were within 95% confidential interval of observed survival in both stage II and III disease, including most demographic and pathologic subgroups. Moreover, AOL more accurately predicted OS for patients with stage II than stage III. AOL tended to offer reliable prediction for 5-year overall survival and could be used as a decision making tool for adjuvant treatment in Korean colon cancer patients whose prognosis is similar to other Asian patients

  8. New probabilistic fracture mechanics approach with neural network-based crack modeling: Its application to multiple cracks problem

    International Nuclear Information System (INIS)

    Studies on efficient utilization and life extension of operating nuclear power plants (NPPs) have become increasingly important since ages of the first-generation NPPs are approaching their design lives. In order to predict a remaining life of each plant, it is necessary to select those critical components that strongly influence the plant life, and to evaluate their remaining lives by considering aging effects of materials and other factors. This paper proposes a new method to incorporate sophisticated crack models, such as interaction and coalescence of multiple surface cracks, into probabilistic fracture mechanism (PFM) computer programs using neural networks. First, hundreds of finite element (FE) calculations of a plate containing multiple surface cracks are performed by parametrically changing crack parameters such as sizes and locations. A fully automated 3D FE analysis system is effectively utilized here. Second, the back-propagation neural network is trained using the FE solutions, i.e. crack parameters vs. their corresponding stress intensity factors (SIFs). After a sufficient number of training iterations, the network attains an ability to promptly output SIFs for arbitrary combinations of crack parameters. The well trained network is then incorporated into the parallel PFM program which runs on one of massively parallel computers composed of 512 processing units. To demonstrate its fundamental performances, the present computer program is applied to evaluate failure probabilities of aged reactor pressure vessels considering interaction and coalescence of two dissimilar semi-elliptical surface cracks

  9. Research and application on integration modeling of 3D bodies in coal mine with blended data model based on TIN and ARTP

    Institute of Scientific and Technical Information of China (English)

    HAN Zuo-zhen; HAN Rui-dong; MAO Shan-jun; HAN Jing-min

    2007-01-01

    Data modeling is the foundation of three-dimensional visualization technology.First the paper proposed the 3D integrated data model of stratum, laneway and drill on the basic of TIN and ARTP, and designed the relevant conceptual and logical model from the view of data model, and described the data structure of geometric elements of the model by adopting the object-oriented modeling idea. And then studied the key modeling technology of stratum, laneway and drill, introduced the ARTP modeling process of stratum,laneway and drill and studied the 3D geometric modeling process of different section laneways. At last, the paper realized the three-dimensional visualization system professionally coalmine-oriented, using SQL Server as background database, Visual C++6.0 and OpenGL as foreground development tools.

  10. DATA MODELING METHOD BASED ON PARTIAL LEAST SQUARE REGRESSION AND APPLICATION IN CORRELATION ANALYSIS OF THE STATOR BARS CONDITION PARAMETERS

    Institute of Scientific and Technical Information of China (English)

    李锐华; 高乃奎; 谢恒堃; 史维祥

    2004-01-01

    Objective To investigate various data message of the stator bars condition parameters under the condition that only a few samples are available, especially about correlation information between the nondestructive parameters and residual breakdown voltage of the stator bars. Methods Artificial stator bars is designed to simulate the generator bars. The partial didcharge( PD) and dielectric loss experiments are performed in order to obtain the nondestructive parameters, and the residual breakdown voltage acquired by AC damage experiment. In order to eliminate the dimension effect on measurement data, raw data is preprocessed by centered-compress. Based on the idea of extracting principal components, a partial least square (PLS) method is applied to screen and synthesize correlation information between the nondestructive parameters and residual breakdown voltage easily. Moreover, various data message about condition parameters are also discussed. Results Graphical analysis function of PLS is easily to understand various data message of the stator bars condition parameters. The analysis Results are consistent with result of aging testing. Conclusion The method can select and extract PLS components of condition parameters from sample data, and the problems of less samples and multicollinearity are solved effectively in regression analysis.

  11. Using Model Driven Engineering technologies for building authoring applications

    OpenAIRE

    Beaudoux, Olivier; Blouin, Arnaud; Jézéquel, Jean-Marc

    2010-01-01

    Building authoring applications is a tedious and complex task that requires a high programming effort. Document technologies, especially XML based ones, can help in reducing such an effort by providing common bases for manipulating documents. Still, the overall task consists mainly of writing the application's source code. Model Driven Engineering (MDE) focuses on generating the source code from an exhaustive model of the application. In this paper, we illustrate that MDE technologies can be ...

  12. Application of an anisotropic bone-remodelling model based on a damage-repair theory to the analysis of the proximal femur before and after total hip replacement.

    Science.gov (United States)

    Doblaré, M; García, J M

    2001-09-01

    In this work, a new model for internal anisotropic bone remodelling is applied to the study of the remodelling behaviour of the proximal femur before and after total hip replacement (THR). This model considers bone remodelling under the scope of a general damage-repair theory following the principles of continuum damage mechanics. A "damage-repair" tensor is defined in terms of the apparent density and Cowin's "fabric tensor", respectively, associated with porosity and directionality of the trabeculae. The different elements of a thermodynamically consistent damage theory are established, including resorption and apposition criteria, evolution law and rate of remodelling. All of these elements were introduced and discussed in detail in a previous paper (García, J. M., Martinez, M. A., Doblaré, M., 2001. An anisotrophic internal-external bone adaptation model based on a combination of CAO and continuum damage mechanics technologies. Computer Methods in Biomechanics and Biomedical Engineering 4(4), 355-378.), including the definition of the proposed mechanical stimulus and the qualitative properties of the model. In this paper, the fundamentals of the proposed model are briefly reviewed and the computational aspects of its implementation are discussed. This model is then applied to the analysis of the remodelling behaviour of the intact femur obtaining densities and mass principal values and directions very close to the experimental data. The second application involved the proximal femoral extremity after THR and the inclusion of an Exeter prosthesis. As a result of the simulation process, some well-known features previously detected in medical clinics were recovered, such as the stress yielding effect in the proximal part of the implant or the enlargement of the cortical layer at the distal part of the implant. With respect to the anisotropic properties, bone microstructure and local stiffness are known to tend to align with the stress principal directions. This

  13. 校园数据集成过渡模式研究与应用%Research and application of data integrated transition model based on campus network

    Institute of Scientific and Technical Information of China (English)

    郭政慧

    2012-01-01

    分析了校园网各业务系统的特点和软件架构模式,提出了利用面向服务架构实现数据集成的过渡方案.采用多粒度服务设计原则,将遗留系统封装成为服务构件.经过比较,选择统一的Web服务标准接口方式,并给出了校园网业务集成的步骤、方法,分析其可行性,并应用于具体实践中.%With the construction of information systems,there has been a wide variety of isolated applications.Based on this,this paper analyzes the characteristics of business systems and software architecture model,then proposes service-oriented architecture for business integration framework.Based on the principles of multi-granularity services,Legacy systems will be encapsulated as a web service component.Finally,the article describes the steps,methods,feasibility of the business integration on compus network.

  14. Crowdsourcing Based 3d Modeling

    Science.gov (United States)

    Somogyi, A.; Barsi, A.; Molnar, B.; Lovas, T.

    2016-06-01

    Web-based photo albums that support organizing and viewing the users' images are widely used. These services provide a convenient solution for storing, editing and sharing images. In many cases, the users attach geotags to the images in order to enable using them e.g. in location based applications on social networks. Our paper discusses a procedure that collects open access images from a site frequently visited by tourists. Geotagged pictures showing the image of a sight or tourist attraction are selected and processed in photogrammetric processing software that produces the 3D model of the captured object. For the particular investigation we selected three attractions in Budapest. To assess the geometrical accuracy, we used laser scanner and DSLR as well as smart phone photography to derive reference values to enable verifying the spatial model obtained from the web-album images. The investigation shows how detailed and accurate models could be derived applying photogrammetric processing software, simply by using images of the community, without visiting the site.

  15. Conceptual Model of User Adaptive Enterprise Application

    Directory of Open Access Journals (Sweden)

    Inese Šūpulniece

    2015-07-01

    Full Text Available The user adaptive enterprise application is a software system, which adapts its behavior to an individual user on the basis of nontrivial inferences from information about the user. The objective of this paper is to elaborate a conceptual model of the user adaptive enterprise applications. In order to conceptualize the user adaptive enterprise applications, their main characteristics are analyzed, the meta-model defining the key concepts relevant to these applications is developed, and the user adaptive enterprise application and its components are defined in terms of the meta-model. Modeling of the user adaptive enterprise application incorporates aspects of enterprise modeling, application modeling, and design of adaptive characteristics of the application. The end-user and her expectations are identified as two concepts of major importance not sufficiently explored in the existing research. Understanding these roles improves the adaptation result in the user adaptive applications.

  16. P-Graph-based Workflow Modelling

    Directory of Open Access Journals (Sweden)

    József Tick

    2007-03-01

    Full Text Available Workflow modelling has been successfully introduced and implemented in severalapplication fields. Therefore, its significance has increased dramatically. Several work flowmodelling techniques have been published so far, out of which quite a number arewidespread applications. For instance the Petri-Net-based modelling has become popularpartly due to its graphical design and partly due to its correct mathematical background.The workflow modelling based on Unified Modelling Language is important because of itspractical usage. This paper introduces and examines the workflow modelling techniquebased on the Process-graph as a possible new solution next to the already existingmodelling techniques.

  17. A Framework for Agile Development of Component-Based Applications

    CERN Document Server

    Waignier, Guillaume; Meur, Anne-Françoise Le; Duchien, Laurence

    2010-01-01

    Agile development processes and component-based software architectures are two software engineering approaches that contribute to enable the rapid building and evolution of applications. Nevertheless, few approaches have proposed a framework to combine agile and component-based development, allowing an application to be tested throughout the entire development cycle. To address this problematic, we have built CALICO, a model-based framework that allows applications to be safely developed in an iterative and incremental manner. The CALICO approach relies on the synchronization of a model view, which specifies the application properties, and a runtime view, which contains the application in its execution context. Tests on the application specifications that require values only known at runtime, are automatically integrated by CALICO into the running application, and the captured needed values are reified at execution time to resume the tests and inform the architect of potential problems. Any modification at th...

  18. LSTM based Conversation Models

    OpenAIRE

    Luan, Yi; Ji, Yangfeng; Ostendorf, Mari

    2016-01-01

    In this paper, we present a conversational model that incorporates both context and participant role for two-party conversations. Different architectures are explored for integrating participant role and context information into a Long Short-term Memory (LSTM) language model. The conversational model can function as a language model or a language generation model. Experiments on the Ubuntu Dialog Corpus show that our model can capture multiple turn interaction between participants. The propos...

  19. Grid-based Meteorological and Crisis Applications

    Science.gov (United States)

    Hluchy, Ladislav; Bartok, Juraj; Tran, Viet; Lucny, Andrej; Gazak, Martin

    2010-05-01

    forecast model is a subject of the parameterization and parameter optimization before its real deployment. The parameter optimization requires tens of evaluations of the parameterized model accuracy and each evaluation of the model parameters requires re-running of the hundreds of meteorological situations collected over the years and comparison of the model output with the observed data. The architecture and inherent heterogeneity of both examples and their computational complexity and their interfaces to other systems and services make them well suited for decomposition into a set of web and grid services. Such decomposition has been performed within several projects we participated or participate in cooperation with academic sphere, namely int.eu.grid (dispersion model deployed as a pilot application to an interactive grid), SEMCO-WS (semantic composition of the web and grid services), DMM (development of a significant meteorological phenomena prediction system based on the data mining), VEGA 2009-2011 and EGEE III. We present useful and practical applications of technologies of high performance computing. The use of grid technology provides access to much higher computation power not only for modeling and simulation, but also for the model parameterization and validation. This results in the model parameters optimization and more accurate simulation outputs. Having taken into account that the simulations are used for the aviation, road traffic and crisis management, even small improvement in accuracy of predictions may result in significant improvement of safety as well as cost reduction. We found grid computing useful for our applications. We are satisfied with this technology and our experience encourages us to extend its use. Within an ongoing project (DMM) we plan to include processing of satellite images which extends our requirement on computation very rapidly. We believe that thanks to grid computing we are able to handle the job almost in real time.

  20. Application patterns for ontology based applications

    OpenAIRE

    Albertsen, Thomas

    2006-01-01

    Software patterns have been proven as a valuable way to storing a repeatable solution to a commonly occurring problem in software design. A pattern is not a finished design that can be directly formed into program code; instead it is a description how to solve a problem that may occur in many situations. In the ontology community very little research have been made in producing high-level patterns where the solution shows how an architecture of an ontology based software might look like. In t...

  1. A New MRI-Based Model of Heart Function with Coupled Hemodynamics and Application to Normal and Diseased Canine Left Ventricles

    OpenAIRE

    Choi, Young Joon; Constantino, Jason; Vedula, Vijay; Trayanova, Natalia; Mittal, Rajat

    2015-01-01

    A methodology for the simulation of heart function that combines an MRI-based model of cardiac electromechanics (CE) with a Navier–Stokes-based hemodynamics model is presented. The CE model consists of two coupled components that simulate the electrical and the mechanical functions of the heart. Accurate representations of ventricular geometry and fiber orientations are constructed from the structural magnetic resonance and the diffusion tensor MR images, respectively. The deformation of the ...

  2. Rigorous model-based uncertainty quantification with application to terminal ballistics—Part II. Systems with uncontrollable inputs and large scatter

    Science.gov (United States)

    Adams, M.; Lashgari, A.; Li, B.; McKerns, M.; Mihaly, J.; Ortiz, M.; Owhadi, H.; Rosakis, A. J.; Stalzer, M.; Sullivan, T. J.

    2012-05-01

    This Part II of this series is concerned with establishing the feasibility of an extended data-on-demand (XDoD) uncertainty quantification (UQ) protocol based on concentration-of-measure inequalities and martingale theory. Specific aims are to establish the feasibility of the protocol and its basic properties, including the tightness of the predictions afforded by the protocol. The assessment is based on an application to terminal ballistics and a specific system configuration consisting of 6061-T6 aluminum plates struck by spherical 440c stainless steel projectiles at ballistic impact speeds in the range of 2.4-2.8 km/s. The system's inputs are the plate thickness, plate obliquity and impact velocity. The perforation area is chosen as the sole performance measure of the system. The objective of the UQ analysis is to certify the lethality of the projectile, i.e., that the projectile perforates the plate with high probability over a prespecified range of impact velocities, plate thicknesses and plate obliquities. All tests were conducted at Caltech's Small Particle Hypervelocity Range (SPHIR), which houses a two-stage gas gun. A feature of this facility is that the impact velocity, while amenable to precise measurement, cannot be controlled precisely but varies randomly according to a known probability density function. In addition, due to a competition between petalling and plugging mechanisms for the material system under consideration, the measured perforation area exhibits considerable scatter. The analysis establishes the feasibility of the XDoD UQ protocol as a rigorous yet practical approach for model-based certification of complex systems characterized by uncontrollable inputs and noisy experimental data.

  3. PBG based terahertz antenna for aerospace applications

    CERN Document Server

    Choudhury, Balamati; Jha, Rakesh Mohan

    2016-01-01

    This book focuses on high-gain antennas in the terahertz spectrum and their optimization. The terahertz spectrum is an unallocated EM spectrum, which is being explored for a number of applications, especially to meet increasing demands of high data rates for wireless space communications. Space communication systems using the terahertz spectrum can resolve the problems of limited bandwidth of present wireless communications without radio-frequency interference. This book describes design of such high-gain antennas and their performance enhancement using photonic band gap (PBG) substrates. Further, optimization of antenna models using evolutionary algorithm based computational engine has been included. The optimized high-performance compact antenna may be used for various wireless applications, such as inter-orbital communications and on-vehicle satellite communications.

  4. Application of Improved Radiation Modeling to General Circulation Models

    Energy Technology Data Exchange (ETDEWEB)

    Michael J Iacono

    2011-04-07

    This research has accomplished its primary objectives of developing accurate and efficient radiation codes, validating them with measurements and higher resolution models, and providing these advancements to the global modeling community to enhance the treatment of cloud and radiative processes in weather and climate prediction models. A critical component of this research has been the development of the longwave and shortwave broadband radiative transfer code for general circulation model (GCM) applications, RRTMG, which is based on the single-column reference code, RRTM, also developed at AER. RRTMG is a rigorously tested radiation model that retains a considerable level of accuracy relative to higher resolution models and measurements despite the performance enhancements that have made it possible to apply this radiation code successfully to global dynamical models. This model includes the radiative effects of all significant atmospheric gases, and it treats the absorption and scattering from liquid and ice clouds and aerosols. RRTMG also includes a statistical technique for representing small-scale cloud variability, such as cloud fraction and the vertical overlap of clouds, which has been shown to improve cloud radiative forcing in global models. This development approach has provided a direct link from observations to the enhanced radiative transfer provided by RRTMG for application to GCMs. Recent comparison of existing climate model radiation codes with high resolution models has documented the improved radiative forcing capability provided by RRTMG, especially at the surface, relative to other GCM radiation models. Due to its high accuracy, its connection to observations, and its computational efficiency, RRTMG has been implemented operationally in many national and international dynamical models to provide validated radiative transfer for improving weather forecasts and enhancing the prediction of global climate change.

  5. Location-based Modeling and Analysis: Tropos-based Approach

    OpenAIRE

    Ali, Raian; Dalpiaz, Fabiano; Giorgini, Paolo

    2008-01-01

    The continuous growth of interest in mobile applications makes the concept of location essential to design and develop software systems. Location-based software is supposed to be able to monitor the location and choose accordingly the most appropriate behavior. In this paper, we propose a novel conceptual framework to model and analyze location-based software. We mainly focus on the social facets of locations adopting concepts such as social actor, resource, and location-based behavior. Our a...

  6. Application of simulation models for the optimization of business processes

    Science.gov (United States)

    Jašek, Roman; Sedláček, Michal; Chramcov, Bronislav; Dvořák, Jiří

    2016-06-01

    The paper deals with the applications of modeling and simulation tools in the optimization of business processes, especially in solving an optimization of signal flow in security company. As a modeling tool was selected Simul8 software that is used to process modeling based on discrete event simulation and which enables the creation of a visual model of production and distribution processes.

  7. Using models to determine irrigation applications for water management

    Science.gov (United States)

    Simple models are used by field researchers and production agriculture to estimate crop water use for the purpose of scheduling irrigation applications. These are generally based on a simple volume balance approach based on estimates of soil water holding capacity, irrigation application amounts, pr...

  8. Holistic, model-based optimization of edge leveling as an enabler for lithographic focus control: application to a memory use case

    Science.gov (United States)

    Hasan, T.; Kang, Y.-S.; Kim, Y.-J.; Park, S.-J.; Jang, S.-Y.; Hu, K.-Y.; Koop, E. J.; Hinnen, P. C.; Voncken, M. M. A. J.

    2016-03-01

    Advancement of the next generation technology nodes and emerging memory devices demand tighter lithographic focus control. Although the leveling performance of the latest-generation scanners is state of the art, challenges remain at the wafer edge due to large process variations. There are several customer configurable leveling control options available in ASML scanners, some of which are application specific in their scope of leveling improvement. In this paper, we assess the usability of leveling non-correctable error models to identify yield limiting edge dies. We introduce a novel dies-inspec based holistic methodology for leveling optimization to guide tool users in selecting an optimal configuration of leveling options. Significant focus gain, and consequently yield gain, can be achieved with this integrated approach. The Samsung site in Hwaseong observed an improved edge focus performance in a production of a mid-end memory product layer running on an ASML NXT 1960 system. 50% improvement in focus and a 1.5%p gain in edge yield were measured with the optimized configurations.

  9. Development of EMC-based empirical model for estimating spatial distribution of pollutant loads and its application in rural areas of Korea.

    Science.gov (United States)

    Yi, Qitao; Li, Hui; Lee, Jin-Woo; Kim, Youngchul

    2015-09-01

    An integrated approach to easily calculate pollutant loads from agricultural watersheds is suggested and verified in this research. The basic concepts of this empirical tool were based on the assumption that variations in event mean concentrations (EMCs) of pollutants from a given agricultural watershed during rainstorms were only attributable to the rainfall pattern. Fifty one sets of EMC values were obtained from nine different watersheds located in the rural areas of Korea, and these data were used to develop predictive tools for the EMCs in rainfall runoff. The results of statistical tests of these formulas show that they are fairly good in predicting actual EMC values of some parameters, and useful in terms of calculating pollutant loads for any rainfall event time span such as daily, weekly, monthly, and yearly. This model was further checked in for its field applicability in a reservoir receiving stormwater after a cleanup of the sediments, covering 17 consecutive rainfall events from 1 July to 15 August in 2007. Overall the predicted values matched the observed values, indicating the feasibility of this empirical tool as a simple and useful solution in evaluating timely distribution of nonpoint source pollution loads from small rural watersheds of Korea. PMID:26354686

  10. Model-based tolerance intervals derived from cumulative historical composition data: application for substantial equivalence assessment of a genetically modified crop.

    Science.gov (United States)

    Hong, Bonnie; Fisher, Tracey L; Sult, Theresa S; Maxwell, Carl A; Mickelson, James A; Kishino, Hirohisa; Locke, Mary E H

    2014-10-01

    Compositional analysis is a requisite component of the substantial equivalence framework utilized to assess genetically modified (GM) crop safety. Statistical differences in composition data between GM and non-GM crops require a context in which to determine biological relevance. This context is provided by surveying the natural variation of key nutrient and antinutrient levels within the crop population with a history of safe use. Data accumulated from various genotypes with a history of safe use cultivated in relevant commercial crop-growing environments over multiple seasons are discussed as the appropriate data representative of this natural variation. A model-based parametric tolerance interval approach, which accounts for the correlated and unbalanced data structure of cumulative historical data collected from multisite field studies conducted over multiple seasons, is presented. This paper promotes the application of this tolerance interval approach to generate reference ranges for evaluation of the biological relevance of statistical differences identified during substantial equivalence assessment of a GM crop. PMID:25208038

  11. Application of measuring 99mTc-MAG3 plasma clearance based on one-compartment model (MPC method) to renal transplantation

    International Nuclear Information System (INIS)

    Measurement of 99mTc-MAG3 plasma clearance (CLmag) based on one-compartment model (MPC method) was applied to renal transplantation and evaluated for the factors which might affect the calculated results, especially concerning renal depth. Correlation coefficient of CLmag between MPC method using real renal depth and Russell or Bubeck single sampling method was good (r=0.852 or 0.876, respectively). Regression equation between MPC method and Russell method was y=1.044x-3.0 and was more closer to y=x than that between MPC method and Bubeck method. CLmag of MPC method calculated by estimated renal depth from the abdominal thickness was also similar to that by real renal depth. Even if the fixed renal depth, 4 cm, was applied, the coefficient and regression equation between MPC method and Russell method were r=0.884 and y=1.004x-10.2. In conclusion, MPC method is applicable to the evaluation of renal transplants. Though measuring renal depth is best, calculation with fixed renal depth of 4 cm might be practically acceptable. (author)

  12. Model-based Abstraction of Data Provenance

    OpenAIRE

    Probst, Christian W.; Hansen, René Rydhof

    2014-01-01

    Identifying provenance of data provides insights to the origin of data and intermediate results, and has recently gained increased interest due to data-centric applications. In this work we extend a data-centric system view with actors handling the data and policies restricting actions. This extension is based on provenance analysis performed on system models. System models have been introduced to model and analyse spatial and organisational aspects of organisations, to identify, e.g., potent...

  13. Multi-variable grey model (MGM (1,n,q)) based on genetic algorithm and its application in urban water consumption

    Institute of Scientific and Technical Information of China (English)

    Yan; Han; Shi; Guoxu

    2007-01-01

    Urban water consumption has some characteristics of grey because it is influenced by economy, population, standard of living and so on. The multi-variable grey model (MGM(1,n)), as the expansion and complement of GM(1,1) model, reveals the relationship between restriction and stimulation among variables, and the genetic algorithm has the whole optimal and parallel characteristics. In this paper, the parameter q of MGM(1,n) model was optimized, and a multi-variable grey model (MGM(1,n,q)) was built by using the genetic algorithm. The model was validated by examining the urban water consumption from 1990 to 2003 in Dalian City. The result indicated that the multi-variable grey model (MGM(1,n,q)) based on genetic algorithm was better than MGM(1,n) model, and the MGM(1,n) model was better than MGM(1,1) model.

  14. Variance Estimation for NAEP Data Using a Resampling-Based Approach: An Application of Cognitive Diagnostic Models. Research Report. ETS RR-10-26

    Science.gov (United States)

    Hsieh, Chueh-an; Xu, Xueli; von Davier, Matthias

    2010-01-01

    This paper presents an application of a jackknifing approach to variance estimation of ability inferences for groups of students, using a multidimensional discrete model for item response data. The data utilized to demonstrate the approach come from the National Assessment of Educational Progress (NAEP). In contrast to the operational approach…

  15. Multiagent-Based Model For ESCM

    OpenAIRE

    Delia MARINCAS

    2011-01-01

    Web based applications for Supply Chain Management (SCM) are now a necessity for every company in order to meet the increasing customer demands, to face the global competition and to make profit. Multiagent-based approach is appropriate for eSCM because it shows many of the characteristics a SCM system should have. For this reason, we have proposed a multiagent-based eSCM model which configures a virtual SC, automates the SC activities: selling, purchasing, manufacturing, planning, inventory,...

  16. Designing Network-based Business Model Ontology

    DEFF Research Database (Denmark)

    Hashemi Nekoo, Ali Reza; Ashourizadeh, Shayegheh; Zarei, Behrouz

    2015-01-01

    is going to propose e-business model ontology from the network point of view and its application in real world. The suggested ontology for network-based businesses is composed of individuals` characteristics and what kind of resources they own. also, their connections and pre-conceptions of connections...... such as shared-mental model and trust. However, it mostly covers previous business model elements. To confirm the applicability of this ontology, it has been implemented in business angel network and showed how it works....

  17. GIS modelling of forest wood residues potential for energy use based on forest inventory data: Methodological approach and case study application

    OpenAIRE

    Panichelli, L.; Gnansounou, E.

    2008-01-01

    This paper presents an approach to perform geo-referenced estimations of forest wood residues availability for energy use based on forest inventory data integration into a GIS. Three different estimation methods are described. The first one evaluates biomass availability based on the application of biomass expansion factors to stem volume data of the forest inventories. The method accounts for forest dynamics and assigns management treatments in function of forest properties. The second metho...

  18. Application of model studies for quality control of bottom pressure based GLOSS sea level gauge at Takoradi Harbour (Ghana, West Africa)

    Digital Repository Service at National Institute of Oceanography (India)

    Joseph, A.; Mehra, P.; Desai, R.G.P.; Dotse, J.; Odammetey, J.T.; Nkebi, E.K.; VijayKumar, K.; Prabhudesai, S.

    Quality-control of bottom pressure based sea level gauge has been effected using a statistically derived simple linear model constructed from a set of bottom pressures and concurrent tide-staff measurements. The study reveals that the crucial factor...

  19. Web协同应用中基于文档划分的一致性维护模型%Document-partition based model for maintaining data consistency in collaborative web-based applications

    Institute of Scientific and Technical Information of China (English)

    陈小碾

    2012-01-01

    The existing approaches for maintaining data consistency in the collaborative web-based applications will result in serious cost for the server. To solve this problem, a document-partition based model for consistency maintenance is proposed. The model introduces the idea of document partitioning on the basis of the operation transformation algorithm SLOT ( symmetric linear operational transformation). From the view of reducing both communication and memory cost, a dynamic document partitioning strategy and corresponding algorithms are proposed to adapt the dynamic of users' behaviors. The experimental result shows that the model can reduce both the communication and memory cost of the server effectively in a large-scale collaborative application.%针对已有的Web协同应用中的一致性维护方法会带来严重的服务器耗费问题,提出了一种基于文档划分的一致性维护模型.该模型在操作转换算法SLOT(symmetric linear operational transformation)的基础上引入文档划分的思想.从降低服务器通信和内存耗费的角度出发,结合用户数量和操作频率的变化,给出一种动态的文档划分策略及其实现算法.仿真实验结果表明,该模型可以有效地降低大规模协同应用中服务器的通信和内存耗费.

  20. System identification application using Hammerstein model

    Indian Academy of Sciences (India)

    SABAN OZER; HASAN ZORLU; SELCUK METE

    2016-06-01

    Generally, memoryless polynomial nonlinear model for nonlinear part and finite impulse response (FIR) model or infinite impulse response model for linear part are preferred in Hammerstein models in literature. In this paper, system identification applications of Hammerstein model that is cascade of nonlinear second order volterra and linear FIR model are studied. Recursive least square algorithm is used to identify the proposed Hammerstein model parameters. Furthermore, the results are compared to identify the success of proposed Hammerstein model and different types of models