WorldWideScience

Sample records for models solution algorithms

  1. A Numerical Algorithm for the Solution of a Phase-Field Model of Polycrystalline Materials

    Energy Technology Data Exchange (ETDEWEB)

    Dorr, M R; Fattebert, J; Wickett, M E; Belak, J F; Turchi, P A

    2008-12-04

    We describe an algorithm for the numerical solution of a phase-field model (PFM) of microstructure evolution in polycrystalline materials. The PFM system of equations includes a local order parameter, a quaternion representation of local orientation and a species composition parameter. The algorithm is based on the implicit integration of a semidiscretization of the PFM system using a backward difference formula (BDF) temporal discretization combined with a Newton-Krylov algorithm to solve the nonlinear system at each time step. The BDF algorithm is combined with a coordinate projection method to maintain quaternion unit length, which is related to an important solution invariant. A key element of the Newton-Krylov algorithm is the selection of a preconditioner to accelerate the convergence of the Generalized Minimum Residual algorithm used to solve the Jacobian linear system in each Newton step. Results are presented for the application of the algorithm to 2D and 3D examples.

  2. Application of Harmony Search algorithm to the solution of groundwater management models

    Science.gov (United States)

    Tamer Ayvaz, M.

    2009-06-01

    This study proposes a groundwater resources management model in which the solution is performed through a combined simulation-optimization model. A modular three-dimensional finite difference groundwater flow model, MODFLOW is used as the simulation model. This model is then combined with a Harmony Search (HS) optimization algorithm which is based on the musical process of searching for a perfect state of harmony. The performance of the proposed HS based management model is tested on three separate groundwater management problems: (i) maximization of total pumping from an aquifer (steady-state); (ii) minimization of the total pumping cost to satisfy the given demand (steady-state); and (iii) minimization of the pumping cost to satisfy the given demand for multiple management periods (transient). The sensitivity of HS algorithm is evaluated by performing a sensitivity analysis which aims to determine the impact of related solution parameters on convergence behavior. The results show that HS yields nearly same or better solutions than the previous solution methods and may be used to solve management problems in groundwater modeling.

  3. An Evaluation of Solution Algorithms and Numerical Approximation Methods for Modeling an Ion Exchange Process.

    Science.gov (United States)

    Bu, Sunyoung; Huang, Jingfang; Boyer, Treavor H; Miller, Cass T

    2010-07-01

    The focus of this work is on the modeling of an ion exchange process that occurs in drinking water treatment applications. The model formulation consists of a two-scale model in which a set of microscale diffusion equations representing ion exchange resin particles that vary in size and age are coupled through a boundary condition with a macroscopic ordinary differential equation (ODE), which represents the concentration of a species in a well-mixed reactor. We introduce a new age-averaged model (AAM) that averages all ion exchange particle ages for a given size particle to avoid the expensive Monte-Carlo simulation associated with previous modeling applications. We discuss two different numerical schemes to approximate both the original Monte Carlo algorithm and the new AAM for this two-scale problem. The first scheme is based on the finite element formulation in space coupled with an existing backward-difference-formula-based ODE solver in time. The second scheme uses an integral equation based Krylov deferred correction (KDC) method and a fast elliptic solver (FES) for the resulting elliptic equations. Numerical results are presented to validate the new AAM algorithm, which is also shown to be more computationally efficient than the original Monte Carlo algorithm. We also demonstrate that the higher order KDC scheme is more efficient than the traditional finite element solution approach and this advantage becomes increasingly important as the desired accuracy of the solution increases. We also discuss issues of smoothness, which affect the efficiency of the KDC-FES approach, and outline additional algorithmic changes that would further improve the efficiency of these developing methods for a wide range of applications.

  4. Solution algorithm of dwell time in slope-based figuring model

    Science.gov (United States)

    Li, Yong; Zhou, Lin

    2017-10-01

    Surface slope profile is commonly used to evaluate X-ray reflective optics, which is used in synchrotron radiation beam. Moreover, the measurement result of measuring instrument for X-ray reflective optics is usually the surface slope profile rather than the surface height profile. To avoid the conversion error, the slope-based figuring model is introduced introduced by processing the X-ray reflective optics based on surface height-based model. However, the pulse iteration method, which can quickly obtain the dell time solution of the traditional height-based figuring model, is not applied to the slope-based figuring model because property of the slope removal function have both positive and negative values and complex asymmetric structure. To overcome this problem, we established the optimal mathematical model for the dwell time solution, By introducing the upper and lower limits of the dwell time and the time gradient constraint. Then we used the constrained least squares algorithm to solve the dwell time in slope-based figuring model. To validate the proposed algorithm, simulations and experiments are conducted. A flat mirror with effective aperture of 80 mm is polished on the ion beam machine. After iterative polishing three times, the surface slope profile error of the workpiece is converged from RMS 5.65 μrad to RMS 1.12 μrad.

  5. Thermodynamically Consistent Algorithms for the Solution of Phase-Field Models

    KAUST Repository

    Vignal, Philippe

    2016-02-11

    Phase-field models are emerging as a promising strategy to simulate interfacial phenomena. Rather than tracking interfaces explicitly as done in sharp interface descriptions, these models use a diffuse order parameter to monitor interfaces implicitly. This implicit description, as well as solid physical and mathematical footings, allow phase-field models to overcome problems found by predecessors. Nonetheless, the method has significant drawbacks. The phase-field framework relies on the solution of high-order, nonlinear partial differential equations. Solving these equations entails a considerable computational cost, so finding efficient strategies to handle them is important. Also, standard discretization strategies can many times lead to incorrect solutions. This happens because, for numerical solutions to phase-field equations to be valid, physical conditions such as mass conservation and free energy monotonicity need to be guaranteed. In this work, we focus on the development of thermodynamically consistent algorithms for time integration of phase-field models. The first part of this thesis focuses on an energy-stable numerical strategy developed for the phase-field crystal equation. This model was put forward to model microstructure evolution. The algorithm developed conserves, guarantees energy stability and is second order accurate in time. The second part of the thesis presents two numerical schemes that generalize literature regarding energy-stable methods for conserved and non-conserved phase-field models. The time discretization strategies can conserve mass if needed, are energy-stable, and second order accurate in time. We also develop an adaptive time-stepping strategy, which can be applied to any second-order accurate scheme. This time-adaptive strategy relies on a backward approximation to give an accurate error estimator. The spatial discretization, in both parts, relies on a mixed finite element formulation and isogeometric analysis. The codes are

  6. Improved Chaff Solution Algorithm

    Science.gov (United States)

    2009-03-01

    Programme de démonstration de technologies (PDT) sur l’intégration de capteurs et de systèmes d’armes embarqués (SISWS), un algorithme a été élaboré...technologies (PDT) sur l’intégration de capteurs et de systèmes d’armes embarqués (SISWS), un algorithme a été élaboré pour déterminer automatiquement

  7. Higher order explicit solutions for nonlinear dynamic model of column buckling using variational approach and variational iteration algorithm-II

    Energy Technology Data Exchange (ETDEWEB)

    Bagheri, Saman; Nikkar, Ali [University of Tabriz, Tabriz (Iran, Islamic Republic of)

    2014-11-15

    This paper deals with the determination of approximate solutions for a model of column buckling using two efficient and powerful methods called He's variational approach and variational iteration algorithm-II. These methods are used to find analytical approximate solution of nonlinear dynamic equation of a model for the column buckling. First and second order approximate solutions of the equation of the system are achieved. To validate the solutions, the analytical results have been compared with those resulted from Runge-Kutta 4th order method. A good agreement of the approximate frequencies and periodic solutions with the numerical results and the exact solution shows that the present methods can be easily extended to other nonlinear oscillation problems in engineering. The accuracy and convenience of the proposed methods are also revealed in comparisons with the other solution techniques.

  8. Model and Algorithm for Substantiating Solutions for Organization of High-Rise Construction Project

    Science.gov (United States)

    Anisimov, Vladimir; Anisimov, Evgeniy; Chernysh, Anatoliy

    2018-03-01

    In the paper the models and the algorithm for the optimal plan formation for the organization of the material and logistical processes of the high-rise construction project and their financial support are developed. The model is based on the representation of the optimization procedure in the form of a non-linear problem of discrete programming, which consists in minimizing the execution time of a set of interrelated works by a limited number of partially interchangeable performers while limiting the total cost of performing the work. The proposed model and algorithm are the basis for creating specific organization management methodologies for the high-rise construction project.

  9. A Solution Approach from an Analytic Model to Heuristic Algorithm for Special Case of Vehicle Routing Problem with Stochastic Demands

    Directory of Open Access Journals (Sweden)

    2009-03-01

    Full Text Available We define a special case for the vehicle routing problem with stochastic demands (SC-VRPSD where customer demands are normally distributed. We propose a new linear model for computing the expected length of a tour in SC-VRPSD. The proposed model is based on the integration of the “Traveling Salesman Problem” (TSP and the Assignment Problem. For large-scale problems, we also use an Iterated Local Search (ILS algorithm in order to reach an effective solution.

  10. Performance of Gradient-Based Solutions versus Genetic Algorithms in the Correlation of Thermal Mathematical Models of Spacecrafts

    Directory of Open Access Journals (Sweden)

    Eva Anglada

    2017-01-01

    Full Text Available The correlation of the thermal mathematical models (TMMs of spacecrafts with the results of the thermal test is a demanding task in terms of time and effort. Theoretically, it can be automatized by means of optimization techniques, although this is a challenging task. Previous studies have shown the ability of genetic algorithms to perform this task in several cases, although some limitations have been detected. In addition, gradient-based methods, although also presenting some limitations, have provided good solutions in other technical fields. For this reason, the performance of genetic algorithms and gradient-based methods in the correlation of TMMs is discussed in this paper to compare the pros and cons of them. The case of study used in the comparison is a real space instrument flown aboard the International Space Station.

  11. Discrete Riccati equation solutions: Distributed algorithms

    Directory of Open Access Journals (Sweden)

    D. G. Lainiotis

    1996-01-01

    Full Text Available In this paper new distributed algorithms for the solution of the discrete Riccati equation are introduced. The algorithms are used to provide robust and computational efficient solutions to the discrete Riccati equation. The proposed distributed algorithms are theoretically interesting and computationally attractive.

  12. Life Cycle Network Modeling Framework and Solution Algorithms for Systems Analysis and Optimization of the Water-Energy Nexus

    Directory of Open Access Journals (Sweden)

    Daniel J. Garcia

    2015-07-01

    Full Text Available The water footprint of energy systems must be considered, as future water scarcity has been identified as a major concern. This work presents a general life cycle network modeling and optimization framework for energy-based products and processes using a functional unit of liters of water consumed in the processing pathway. We analyze and optimize the water-energy nexus over the objectives of water footprint minimization, maximization of economic output per liter of water consumed (economic efficiency of water, and maximization of energy output per liter of water consumed (energy efficiency of water. A mixed integer, multiobjective nonlinear fractional programming (MINLFP model is formulated. A mixed integer linear programing (MILP-based branch and refine algorithm that incorporates both the parametric algorithm and nonlinear programming (NLP subproblems is developed to boost solving efficiency. A case study in bioenergy is presented, and the water footprint is considered from biomass cultivation to biofuel production, providing a novel perspective into the consumption of water throughout the value chain. The case study, optimized successively over the three aforementioned objectives, utilizes a variety of candidate biomass feedstocks to meet primary fuel products demand (ethanol, diesel, and gasoline. A minimum water footprint of 55.1 ML/year was found, economic efficiencies of water range from −$1.31/L to $0.76/L, and energy efficiencies of water ranged from 15.32 MJ/L to 27.98 MJ/L. These results show optimization provides avenues for process improvement, as reported values for the energy efficiency of bioethanol range from 0.62 MJ/L to 3.18 MJ/L. Furthermore, the proposed solution approach was shown to be an order of magnitude more efficient than directly solving the original MINLFP problem with general purpose solvers.

  13. A new algorithm for anisotropic solutions

    Indian Academy of Sciences (India)

    We establish a new algorithm that generates a new solution to the Einstein field equations, with an anisotropic matter distribution, from a seed isotropic solution. The new solution is expressed in terms of integrals of an isotropic gravitational potential; and the integration can be completed exactly for particular isotropic seed ...

  14. Cloud Model Bat Algorithm

    OpenAIRE

    Yongquan Zhou; Jian Xie; Liangliang Li; Mingzhi Ma

    2014-01-01

    Bat algorithm (BA) is a novel stochastic global optimization algorithm. Cloud model is an effective tool in transforming between qualitative concepts and their quantitative representation. Based on the bat echolocation mechanism and excellent characteristics of cloud model on uncertainty knowledge representation, a new cloud model bat algorithm (CBA) is proposed. This paper focuses on remodeling echolocation model based on living and preying characteristics of bats, utilizing the transformati...

  15. Cloud model bat algorithm.

    Science.gov (United States)

    Zhou, Yongquan; Xie, Jian; Li, Liangliang; Ma, Mingzhi

    2014-01-01

    Bat algorithm (BA) is a novel stochastic global optimization algorithm. Cloud model is an effective tool in transforming between qualitative concepts and their quantitative representation. Based on the bat echolocation mechanism and excellent characteristics of cloud model on uncertainty knowledge representation, a new cloud model bat algorithm (CBA) is proposed. This paper focuses on remodeling echolocation model based on living and preying characteristics of bats, utilizing the transformation theory of cloud model to depict the qualitative concept: "bats approach their prey." Furthermore, Lévy flight mode and population information communication mechanism of bats are introduced to balance the advantage between exploration and exploitation. The simulation results show that the cloud model bat algorithm has good performance on functions optimization.

  16. Cloud Model Bat Algorithm

    Directory of Open Access Journals (Sweden)

    Yongquan Zhou

    2014-01-01

    Full Text Available Bat algorithm (BA is a novel stochastic global optimization algorithm. Cloud model is an effective tool in transforming between qualitative concepts and their quantitative representation. Based on the bat echolocation mechanism and excellent characteristics of cloud model on uncertainty knowledge representation, a new cloud model bat algorithm (CBA is proposed. This paper focuses on remodeling echolocation model based on living and preying characteristics of bats, utilizing the transformation theory of cloud model to depict the qualitative concept: “bats approach their prey.” Furthermore, Lévy flight mode and population information communication mechanism of bats are introduced to balance the advantage between exploration and exploitation. The simulation results show that the cloud model bat algorithm has good performance on functions optimization.

  17. Modeling Design Iteration in Product Design and Development and Its Solution by a Novel Artificial Bee Colony Algorithm

    Science.gov (United States)

    2014-01-01

    Due to fierce market competition, how to improve product quality and reduce development cost determines the core competitiveness of enterprises. However, design iteration generally causes increases of product cost and delays of development time as well, so how to identify and model couplings among tasks in product design and development has become an important issue for enterprises to settle. In this paper, the shortcomings existing in WTM model are discussed and tearing approach as well as inner iteration method is used to complement the classic WTM model. In addition, the ABC algorithm is also introduced to find out the optimal decoupling schemes. In this paper, firstly, tearing approach and inner iteration method are analyzed for solving coupled sets. Secondly, a hybrid iteration model combining these two technologies is set up. Thirdly, a high-performance swarm intelligence algorithm, artificial bee colony, is adopted to realize problem-solving. Finally, an engineering design of a chemical processing system is given in order to verify its reasonability and effectiveness. PMID:25431584

  18. Complex fluids modeling and algorithms

    CERN Document Server

    Saramito, Pierre

    2016-01-01

    This book presents a comprehensive overview of the modeling of complex fluids, including many common substances, such as toothpaste, hair gel, mayonnaise, liquid foam, cement and blood, which cannot be described by Navier-Stokes equations. It also offers an up-to-date mathematical and numerical analysis of the corresponding equations, as well as several practical numerical algorithms and software solutions for the approximation of the solutions. It discusses industrial (molten plastics, forming process), geophysical (mud flows, volcanic lava, glaciers and snow avalanches), and biological (blood flows, tissues) modeling applications. This book is a valuable resource for undergraduate students and researchers in applied mathematics, mechanical engineering and physics.

  19. A new algorithm for anisotropic solutions

    Indian Academy of Sciences (India)

    Introduction. Numerous models of static perfect fluid spheres, in the context of general relativity, have been constructed in the past because these are first approximations in building a realistic model for a star. Lists of exact solutions to the Einstein field equations modelling relativistic perfect fluid spheres are given in several ...

  20. On the multi-level solution algorithm for Markov chains

    Energy Technology Data Exchange (ETDEWEB)

    Horton, G. [Univ. of Erlangen, Nuernberg (Germany)

    1996-12-31

    We discuss the recently introduced multi-level algorithm for the steady-state solution of Markov chains. The method is based on the aggregation principle, which is well established in the literature. Recursive application of the aggregation yields a multi-level method which has been shown experimentally to give results significantly faster than the methods currently in use. The algorithm can be reformulated as an algebraic multigrid scheme of Galerkin-full approximation type. The uniqueness of the scheme stems from its solution-dependent prolongation operator which permits significant computational savings in the evaluation of certain terms. This paper describes the modeling of computer systems to derive information on performance, measured typically as job throughput or component utilization, and availability, defined as the proportion of time a system is able to perform a certain function in the presence of component failures and possibly also repairs.

  1. Solutions of the Two-Dimensional Hubbard Model: Benchmarks and Results from a Wide Range of Numerical Algorithms

    Directory of Open Access Journals (Sweden)

    2015-12-01

    Full Text Available Numerical results for ground-state and excited-state properties (energies, double occupancies, and Matsubara-axis self-energies of the single-orbital Hubbard model on a two-dimensional square lattice are presented, in order to provide an assessment of our ability to compute accurate results in the thermodynamic limit. Many methods are employed, including auxiliary-field quantum Monte Carlo, bare and bold-line diagrammatic Monte Carlo, method of dual fermions, density matrix embedding theory, density matrix renormalization group, dynamical cluster approximation, diffusion Monte Carlo within a fixed-node approximation, unrestricted coupled cluster theory, and multireference projected Hartree-Fock methods. Comparison of results obtained by different methods allows for the identification of uncertainties and systematic errors. The importance of extrapolation to converged thermodynamic-limit values is emphasized. Cases where agreement between different methods is obtained establish benchmark results that may be useful in the validation of new approaches and the improvement of existing methods.

  2. A discretized algorithm for the solution of a constrained, continuous ...

    African Journals Online (AJOL)

    A discretized algorithm for the solution of a constrained, continuous quadratic control problem. ... Numerical solution techniques such as Function space algorithm (FSA), Extended conjugate gradient method (ECGM) and Imbedding extended conjugate gradient method (MECGM) are common techniques for solving optimal ...

  3. Genetic algorithm solution for partial digest problem.

    Science.gov (United States)

    Ahrabian, Hayedeh; Ganjtabesh, Mohammad; Nowzari-Dalini, Abbas; Razaghi-Moghadam-Kashani, Zahra

    2013-01-01

    One of the fundamental problems in computational biology is the construction of physical maps of chromosomes from the hybridisation experiments between unique probes and clones of chromosome fragments. Before introducing the shotgun sequencing method, Partial Digest Problem (PDP) was an intractable problem used to construct the physical maps of DNA sequence in molecular biology. In this paper, we develop a novel Genetic Algorithm (GA) for solving the PDP. This algorithm is implemented and compared with well-known existing algorithms on different types of random and real instances data, and the obtained results show the efficiency of our algorithm. Also, our GA is adapted to handle the erroneous data and their efficiency is presented for the large instances of this problem.

  4. Finite element solution algorithm for incompressible fluid dynamics

    Science.gov (United States)

    Baker, A. J.

    1974-01-01

    A finite element solution algorithm is established for the two-dimensional Navier-Stokes equations governing the transient motion of a viscous incompressible fluid, i.e., hydrodynamics. Dependent variable transformation renders the differential equation description uniformly elliptic. The finite element algorithm is established using the Galerkin criterion on a local basis within the Method of Weighted Residuals. It is unconstrained with respect to system linearity, computational mesh uniformity or solution domain closure regularity. The finite element matrices are established using a linear 'natural coordinate function' description. Computational solutions using the COMOC computer program illustrate the various features of the algorithm including recirculating flows.

  5. a permutation encoding te algorithm solution of reso tation encoding

    African Journals Online (AJOL)

    eobe

    even projects. Resource constrained pro scheduling problems (RCPSPS) involve assigning or tasks to a resource or set of resources with lim capacity in order to .... GENETIC ALGORITHM SOLUTION OF RESOURCE CONSTRAINED PROJECT SCHEDULING PROBLEM ..... Journal of Advanced Computer Science and.

  6. Massively Parallel Algorithms for Solution of Schrodinger Equation

    Science.gov (United States)

    Fijany, Amir; Barhen, Jacob; Toomerian, Nikzad

    1994-01-01

    In this paper massively parallel algorithms for solution of Schrodinger equation are developed. Our results clearly indicate that the Crank-Nicolson method, in addition to its excellent numerical properties, is also highly suitable for massively parallel computation.

  7. SOLUTION OF THE SATELLITE TRANSFER PROBLEM WITH HYBRID MEMETIC ALGORITHM

    Directory of Open Access Journals (Sweden)

    A. V. Panteleyev

    2014-01-01

    Full Text Available This paper presents a hybrid memetic algorithm (MA to solve the problem of finding the optimal program control of nonlinear continuous deterministic systems based on the concept of the meme, which is one of the promising solutions obtained in the course of implementing the procedure for searching the extremes. On the basis of the proposed algorithm the software complex is formed in C#. The solution of satellite transfer problem is presented.

  8. Multiagent scheduling models and algorithms

    CERN Document Server

    Agnetis, Alessandro; Gawiejnowicz, Stanisław; Pacciarelli, Dario; Soukhal, Ameur

    2014-01-01

    This book presents multi-agent scheduling models in which subsets of jobs sharing the same resources are evaluated by different criteria. It discusses complexity results, approximation schemes, heuristics and exact algorithms.

  9. A genetic algorithm for solving supply chain network design model

    Science.gov (United States)

    Firoozi, Z.; Ismail, N.; Ariafar, S. H.; Tang, S. H.; Ariffin, M. K. M. A.

    2013-09-01

    Network design is by nature costly and optimization models play significant role in reducing the unnecessary cost components of a distribution network. This study proposes a genetic algorithm to solve a distribution network design model. The structure of the chromosome in the proposed algorithm is defined in a novel way that in addition to producing feasible solutions, it also reduces the computational complexity of the algorithm. Computational results are presented to show the algorithm performance.

  10. Accurate solution algorithms for incompressible multiphase flows

    International Nuclear Information System (INIS)

    Rider, W.J.; Kothe, D.B.; Mosso, S.J.; Cerutti, J.H.; Hochstein, J.I.

    1994-01-01

    A number of advances in modeling multiphase incompressible flow are described. These advances include high-order Godunov projection methods, piecewise linear interface reconstruction and tracking and the continuum surface force model. Examples are given

  11. Parallel Algorithms for Model Checking

    NARCIS (Netherlands)

    van de Pol, Jaco; Mousavi, Mohammad Reza; Sgall, Jiri

    2017-01-01

    Model checking is an automated verification procedure, which checks that a model of a system satisfies certain properties. These properties are typically expressed in some temporal logic, like LTL and CTL. Algorithms for LTL model checking (linear time logic) are based on automata theory and graph

  12. A solution quality assessment method for swarm intelligence optimization algorithms.

    Science.gov (United States)

    Zhang, Zhaojun; Wang, Gai-Ge; Zou, Kuansheng; Zhang, Jianhua

    2014-01-01

    Nowadays, swarm intelligence optimization has become an important optimization tool and wildly used in many fields of application. In contrast to many successful applications, the theoretical foundation is rather weak. Therefore, there are still many problems to be solved. One problem is how to quantify the performance of algorithm in finite time, that is, how to evaluate the solution quality got by algorithm for practical problems. It greatly limits the application in practical problems. A solution quality assessment method for intelligent optimization is proposed in this paper. It is an experimental analysis method based on the analysis of search space and characteristic of algorithm itself. Instead of "value performance," the "ordinal performance" is used as evaluation criteria in this method. The feasible solutions were clustered according to distance to divide solution samples into several parts. Then, solution space and "good enough" set can be decomposed based on the clustering results. Last, using relative knowledge of statistics, the evaluation result can be got. To validate the proposed method, some intelligent algorithms such as ant colony optimization (ACO), particle swarm optimization (PSO), and artificial fish swarm algorithm (AFS) were taken to solve traveling salesman problem. Computational results indicate the feasibility of proposed method.

  13. A Solution Quality Assessment Method for Swarm Intelligence Optimization Algorithms

    Directory of Open Access Journals (Sweden)

    Zhaojun Zhang

    2014-01-01

    Full Text Available Nowadays, swarm intelligence optimization has become an important optimization tool and wildly used in many fields of application. In contrast to many successful applications, the theoretical foundation is rather weak. Therefore, there are still many problems to be solved. One problem is how to quantify the performance of algorithm in finite time, that is, how to evaluate the solution quality got by algorithm for practical problems. It greatly limits the application in practical problems. A solution quality assessment method for intelligent optimization is proposed in this paper. It is an experimental analysis method based on the analysis of search space and characteristic of algorithm itself. Instead of “value performance,” the “ordinal performance” is used as evaluation criteria in this method. The feasible solutions were clustered according to distance to divide solution samples into several parts. Then, solution space and “good enough” set can be decomposed based on the clustering results. Last, using relative knowledge of statistics, the evaluation result can be got. To validate the proposed method, some intelligent algorithms such as ant colony optimization (ACO, particle swarm optimization (PSO, and artificial fish swarm algorithm (AFS were taken to solve traveling salesman problem. Computational results indicate the feasibility of proposed method.

  14. Applied Integer Programming Modeling and Solution

    CERN Document Server

    Chen, Der-San; Dang, Yu

    2011-01-01

    An accessible treatment of the modeling and solution of integer programming problems, featuring modern applications and software In order to fully comprehend the algorithms associated with integer programming, it is important to understand not only how algorithms work, but also why they work. Applied Integer Programming features a unique emphasis on this point, focusing on problem modeling and solution using commercial software. Taking an application-oriented approach, this book addresses the art and science of mathematical modeling related to the mixed integer programming (MIP) framework and

  15. Modeling of Nonlinear Systems using Genetic Algorithm

    Science.gov (United States)

    Hayashi, Kayoko; Yamamoto, Toru; Kawada, Kazuo

    In this paper, a newly modeling system by using Genetic Algorithm (GA) is proposed. The GA is an evolutionary computational method that simulates the mechanisms of heredity or evolution of living things, and it is utilized in optimization and in searching for optimized solutions. Most process systems have nonlinearities, so it is necessary to anticipate exactly such systems. However, it is difficult to make a suitable model for nonlinear systems, because most nonlinear systems have a complex structure. Therefore the newly proposed method of modeling for nonlinear systems uses GA. Then, according to the newly proposed scheme, the optimal structure and parameters of the nonlinear model are automatically generated.

  16. Algorithmic Issues in Modeling Motion

    DEFF Research Database (Denmark)

    Agarwal, P. K; Guibas, L. J; Edelsbrunner, H.

    2003-01-01

    This article is a survey of research areas in which motion plays a pivotal role. The aim of the article is to review current approaches to modeling motion together with related data structures and algorithms, and to summarize the challenges that lie ahead in producing a more unified theory...

  17. Algorithmic solution of arithmetic problems and operands-answer associations in long-term memory.

    Science.gov (United States)

    Thevenot, C; Barrouillet, P; Fayol, M

    2001-05-01

    Many developmental models of arithmetic problem solving assume that any algorithmic solution of a given problem results in an association of the two operands and the answer in memory (Logan & Klapp, 1991; Siegler, 1996). In this experiment, adults had to perform either an operation or a comparison on the same pairs of two-digit numbers and then a recognition task. It is shown that unlike comparisons, the algorithmic solution of operations impairs the recognition of operands in adults. Thus, the postulate of a necessary and automatic storage of operands-answer associations in memory when young children solve additions by algorithmic strategies needs to be qualified.

  18. The linear ordering problem: an algorithm for the optimal solution ...

    African Journals Online (AJOL)

    In this paper we describe and implement an algorithm for the exact solution of the Linear Ordering problem. Linear Ordering is the problem of finding a linear order of the nodes of a graph such that the sum of the weights which are consistent with this order is as large as possible. It is an NP - Hard combinatorial optimisation ...

  19. Autonomous path planning solution for industrial robot manipulator using backpropagation algorithm

    Directory of Open Access Journals (Sweden)

    PeiJiang Yuan

    2015-12-01

    Full Text Available Here, we propose an autonomous path planning solution using backpropagation algorithm. The mechanism of movement used by humans in controlling their arms is analyzed and then applied to control a robot manipulator. Autonomous path planning solution is a numerical method. The model of industrial robot manipulator used in this article is a KUKA KR 210 R2700 EXTRA robot. In order to show the performance of the autonomous path planning solution, an experiment validation of path tracking is provided. Experiment validation consists of implementation of the autonomous path planning solution and the control of physical robot. The process of converging to target solution is provided. The mean absolute error of position for tool center point is also analyzed. Comparison between autonomous path planning solution and the numerical methods based on Newton–Raphson algorithm is provided to demonstrate the efficiency and accuracy of the autonomous path planning solution.

  20. Models and algorithms for biomolecules and molecular networks

    CERN Document Server

    DasGupta, Bhaskar

    2016-01-01

    By providing expositions to modeling principles, theories, computational solutions, and open problems, this reference presents a full scope on relevant biological phenomena, modeling frameworks, technical challenges, and algorithms. * Up-to-date developments of structures of biomolecules, systems biology, advanced models, and algorithms * Sampling techniques for estimating evolutionary rates and generating molecular structures * Accurate computation of probability landscape of stochastic networks, solving discrete chemical master equations * End-of-chapter exercises

  1. Fireworks algorithm for mean-VaR/CVaR models

    Science.gov (United States)

    Zhang, Tingting; Liu, Zhifeng

    2017-10-01

    Intelligent algorithms have been widely applied to portfolio optimization problems. In this paper, we introduce a novel intelligent algorithm, named fireworks algorithm, to solve the mean-VaR/CVaR model for the first time. The results show that, compared with the classical genetic algorithm, fireworks algorithm not only improves the optimization accuracy and the optimization speed, but also makes the optimal solution more stable. We repeat our experiments at different confidence levels and different degrees of risk aversion, and the results are robust. It suggests that fireworks algorithm has more advantages than genetic algorithm in solving the portfolio optimization problem, and it is feasible and promising to apply it into this field.

  2. A Solution Generator Algorithm for Decision Making based Automated Negotiation in the Construction Domain

    Directory of Open Access Journals (Sweden)

    Arazi Idrus

    2017-12-01

    Full Text Available In this paper, we present our work-in-progress of a proposed framework for automated negotiation in the construction domain. The proposed framework enables software agents to conduct negotiations and autonomously make value-based decisions. The framework consists of three main components which are, solution generator algorithm, negotiation algorithm, and conflict resolution algorithm. This paper extends the discussion on the solution generator algorithm that enables software agents to generate solutions and rank them from 1st to nth solution for the negotiation stage of the operation. The solution generator algorithm consists of three steps which are, review solutions, rank solutions, and form ranked solutions. For validation purpose, we present a scenario that utilizes the proposed algorithm to rank solutions. The validation shows that the algorithm is promising, however, it also highlights the conflict between different parties that needs further negotiation action.

  3. Information Dynamics in Networks: Models and Algorithms

    Science.gov (United States)

    2016-09-13

    Information Dynamics in Networks: Models and Algorithms In this project, we investigated how network structure interplays with higher level processes in...Models and Algorithms Report Title In this project, we investigated how network structure interplays with higher level processes in online social...Received Paper 1.00 2.00 3.00 . A Note on Modeling Retweet Cascades on Twitter, Workshop on Algorithms and Models for the Web Graph. 09-DEC-15

  4. Static Load Balancing Algorithms In Cloud Computing Challenges amp Solutions

    Directory of Open Access Journals (Sweden)

    Nadeem Shah

    2015-08-01

    Full Text Available Abstract Cloud computing provides on-demand hosted computing resources and services over the Internet on a pay-per-use basis. It is currently becoming the favored method of communication and computation over scalable networks due to numerous attractive attributes such as high availability scalability fault tolerance simplicity of management and low cost of ownership. Due to the huge demand of cloud computing efficient load balancing becomes critical to ensure that computational tasks are evenly distributed across servers to prevent bottlenecks. The aim of this review paper is to understand the current challenges in cloud computing primarily in cloud load balancing using static algorithms and finding gaps to bridge for more efficient static cloud load balancing in the future. We believe the ideas suggested as new solution will allow researchers to redesign better algorithms for better functionalities and improved user experiences in simple cloud systems. This could assist small businesses that cannot afford infrastructure that supports complex amp dynamic load balancing algorithms.

  5. Artificial Neural Network Modeling and Genetic Algorithm Optimization for Cadmium Removal from Aqueous Solutions by Reduced Graphene Oxide-Supported Nanoscale Zero-Valent Iron (nZVI/rGO) Composites.

    Science.gov (United States)

    Fan, Mingyi; Li, Tongjun; Hu, Jiwei; Cao, Rensheng; Wei, Xionghui; Shi, Xuedan; Ruan, Wenqian

    2017-05-17

    Reduced graphene oxide-supported nanoscale zero-valent iron (nZVI/rGO) composites were synthesized in the present study by chemical deposition method and were then characterized by various methods, such as Fourier-transform infrared spectroscopy (FTIR) and X-ray photoelectron spectroscopy (XPS). The nZVI/rGO composites prepared were utilized for Cd(II) removal from aqueous solutions in batch mode at different initial Cd(II) concentrations, initial pH values, contact times, and operating temperatures. Response surface methodology (RSM) and artificial neural network hybridized with genetic algorithm (ANN-GA) were used for modeling the removal efficiency of Cd(II) and optimizing the four removal process variables. The average values of prediction errors for the RSM and ANN-GA models were 6.47% and 1.08%. Although both models were proven to be reliable in terms of predicting the removal efficiency of Cd(II), the ANN-GA model was found to be more accurate than the RSM model. In addition, experimental data were fitted to the Langmuir, Freundlich, and Dubinin-Radushkevich (D-R) isotherms. It was found that the Cd(II) adsorption was best fitted to the Langmuir isotherm. Examination on thermodynamic parameters revealed that the removal process was spontaneous and exothermic in nature. Furthermore, the pseudo-second-order model can better describe the kinetics of Cd(II) removal with a good R² value than the pseudo-first-order model.

  6. Warehouse Optimization Model Based on Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Guofeng Qin

    2013-01-01

    Full Text Available This paper takes Bao Steel logistics automated warehouse system as an example. The premise is to maintain the focus of the shelf below half of the height of the shelf. As a result, the cost time of getting or putting goods on the shelf is reduced, and the distance of the same kind of goods is also reduced. Construct a multiobjective optimization model, using genetic algorithm to optimize problem. At last, we get a local optimal solution. Before optimization, the average cost time of getting or putting goods is 4.52996 s, and the average distance of the same kinds of goods is 2.35318 m. After optimization, the average cost time is 4.28859 s, and the average distance is 1.97366 m. After analysis, we can draw the conclusion that this model can improve the efficiency of cargo storage.

  7. Feedback model predictive control by randomized algorithms

    NARCIS (Netherlands)

    Batina, Ivo; Stoorvogel, Antonie Arij; Weiland, Siep

    2001-01-01

    In this paper we present a further development of an algorithm for stochastic disturbance rejection in model predictive control with input constraints based on randomized algorithms. The algorithm presented in our work can solve the problem of stochastic disturbance rejection approximately but with

  8. A Robustly Stabilizing Model Predictive Control Algorithm

    Science.gov (United States)

    Ackmece, A. Behcet; Carson, John M., III

    2007-01-01

    A model predictive control (MPC) algorithm that differs from prior MPC algorithms has been developed for controlling an uncertain nonlinear system. This algorithm guarantees the resolvability of an associated finite-horizon optimal-control problem in a receding-horizon implementation.

  9. ALGORITHM OF SELECTION EFFECTIVE SOLUTIONS FOR REPROFILING OF INDUSTRIAL BUILDINGS

    Directory of Open Access Journals (Sweden)

    MENEJLJUK A. I.

    2016-08-01

    Full Text Available Raising of problem.Non-compliance requirements of today's industrial enterprises, which were built during the Soviet period, as well as significant technical progress, economic reform and transition to market principles of performance evaluation leading to necessity to change their target and functionality. The technical condition of many industrial buildings in Ukraine allows to exploit them for decades.Redesigning manufacturing enterprises allows not only to reduce the cost of construction, but also to obtain new facilities in the city. Despite the large number of industrial buildings that have lost their effectiveness and relevance, as well as a significant investor interest in these objects, the scope of redevelopment in the construction remains unexplored. Analysis researches on the topic. The problem of reconstruction of industrial buildings considered in Topchy D. [3], Travin V. [9], as well as in the work of other scientists. However, there are no rules in regulatory documents and system studies for improving the organization of the reconstruction of buildings at realigning. The purpose of this work is the development an algorithm of actions for selection of effective organizational decisions at the planning stage of a reprofiling project of industrial buildings. The proposed algorithm allows you to select an effective organizational and technological solution for the re-profiling of industrial buildings, taking into account features of the building, its location, its state of structures and existing restrictions. The most effective organizational solution allows realize the reprofiling project of an industrial building in the most possible short terms and with the lowest possible use of material resources, taking into account the available features and restrictions. Conclusion. Each object has a number of unique features that necessary for considering at choosing an effective reprofiling variant. The developed algorithm for selecting

  10. Modeling and Engineering Algorithms for Mobile Data

    DEFF Research Database (Denmark)

    Blunck, Henrik; Hinrichs, Klaus; Sondern, Joëlle

    2006-01-01

    In this paper, we present an object-oriented approach to modeling mobile data and algorithms operating on such data. Our model is general enough to capture any kind of continuous motion while at the same time allowing for encompassing algorithms optimized for specific types of motion. Such motion...

  11. Algorithms to solve the Sutherland model

    OpenAIRE

    Langmann, Edwin

    2001-01-01

    We give a self-contained presentation and comparison of two different algorithms to explicitly solve quantum many body models of indistinguishable particles moving on a circle and interacting with two-body potentials of $1/\\sin^2$-type. The first algorithm is due to Sutherland and well-known; the second one is a limiting case of a novel algorithm to solve the elliptic generalization of the Sutherland model. These two algorithms are different in several details. We show that they are equivalen...

  12. DiamondTorre Algorithm for High-Performance Wave Modeling

    Directory of Open Access Journals (Sweden)

    Vadim Levchenko

    2016-08-01

    Full Text Available Effective algorithms of physical media numerical modeling problems’ solution are discussed. The computation rate of such problems is limited by memory bandwidth if implemented with traditional algorithms. The numerical solution of the wave equation is considered. A finite difference scheme with a cross stencil and a high order of approximation is used. The DiamondTorre algorithm is constructed, with regard to the specifics of the GPGPU’s (general purpose graphical processing unit memory hierarchy and parallelism. The advantages of these algorithms are a high level of data localization, as well as the property of asynchrony, which allows one to effectively utilize all levels of GPGPU parallelism. The computational intensity of the algorithm is greater than the one for the best traditional algorithms with stepwise synchronization. As a consequence, it becomes possible to overcome the above-mentioned limitation. The algorithm is implemented with CUDA. For the scheme with the second order of approximation, the calculation performance of 50 billion cells per second is achieved. This exceeds the result of the best traditional algorithm by a factor of five.

  13. Artificial Neural Network Modeling and Genetic Algorithm Optimization for Cadmium Removal from Aqueous Solutions by Reduced Graphene Oxide-Supported Nanoscale Zero-Valent Iron (nZVI/rGO Composites

    Directory of Open Access Journals (Sweden)

    Mingyi Fan

    2017-05-01

    Full Text Available Reduced graphene oxide-supported nanoscale zero-valent iron (nZVI/rGO composites were synthesized in the present study by chemical deposition method and were then characterized by various methods, such as Fourier-transform infrared spectroscopy (FTIR and X-ray photoelectron spectroscopy (XPS. The nZVI/rGO composites prepared were utilized for Cd(II removal from aqueous solutions in batch mode at different initial Cd(II concentrations, initial pH values, contact times, and operating temperatures. Response surface methodology (RSM and artificial neural network hybridized with genetic algorithm (ANN-GA were used for modeling the removal efficiency of Cd(II and optimizing the four removal process variables. The average values of prediction errors for the RSM and ANN-GA models were 6.47% and 1.08%. Although both models were proven to be reliable in terms of predicting the removal efficiency of Cd(II, the ANN-GA model was found to be more accurate than the RSM model. In addition, experimental data were fitted to the Langmuir, Freundlich, and Dubinin-Radushkevich (D-R isotherms. It was found that the Cd(II adsorption was best fitted to the Langmuir isotherm. Examination on thermodynamic parameters revealed that the removal process was spontaneous and exothermic in nature. Furthermore, the pseudo-second-order model can better describe the kinetics of Cd(II removal with a good R2 value than the pseudo-first-order model.

  14. Quantitative Methods in Supply Chain Management Models and Algorithms

    CERN Document Server

    Christou, Ioannis T

    2012-01-01

    Quantitative Methods in Supply Chain Management presents some of the most important methods and tools available for modeling and solving problems arising in the context of supply chain management. In the context of this book, “solving problems” usually means designing efficient algorithms for obtaining high-quality solutions. The first chapter is an extensive optimization review covering continuous unconstrained and constrained linear and nonlinear optimization algorithms, as well as dynamic programming and discrete optimization exact methods and heuristics. The second chapter presents time-series forecasting methods together with prediction market techniques for demand forecasting of new products and services. The third chapter details models and algorithms for planning and scheduling with an emphasis on production planning and personnel scheduling. The fourth chapter presents deterministic and stochastic models for inventory control with a detailed analysis on periodic review systems and algorithmic dev...

  15. How to incorporate generic refraction models into multistatic tracking algorithms

    Science.gov (United States)

    Crouse, D. F.

    The vast majority of literature published on target tracking ignores the effects of atmospheric refraction. When refraction is considered, the solutions are generally tailored to a simple exponential atmospheric refraction model. This paper discusses how arbitrary refraction models can be incorporated into tracking algorithms. Attention is paid to multistatic tracking problems, where uncorrected refractive effects can worsen track accuracy and consistency in centralized tracking algorithms, and can lead to difficulties in track-to-track association in distributed tracking filters. Monostatic and bistatic track initialization using refraction-corrupted measurements is discussed. The results are demonstrated using an exponential refractive model, though an arbitrary refraction profile can be substituted.

  16. LCD motion blur: modeling, analysis, and algorithm.

    Science.gov (United States)

    Chan, Stanley H; Nguyen, Truong Q

    2011-08-01

    Liquid crystal display (LCD) devices are well known for their slow responses due to the physical limitations of liquid crystals. Therefore, fast moving objects in a scene are often perceived as blurred. This effect is known as the LCD motion blur. In order to reduce LCD motion blur, an accurate LCD model and an efficient deblurring algorithm are needed. However, existing LCD motion blur models are insufficient to reflect the limitation of human-eye-tracking system. Also, the spatiotemporal equivalence in LCD motion blur models has not been proven directly in the discrete 2-D spatial domain, although it is widely used. There are three main contributions of this paper: modeling, analysis, and algorithm. First, a comprehensive LCD motion blur model is presented, in which human-eye-tracking limits are taken into consideration. Second, a complete analysis of spatiotemporal equivalence is provided and verified using real video sequences. Third, an LCD motion blur reduction algorithm is proposed. The proposed algorithm solves an l(1)-norm regularized least-squares minimization problem using a subgradient projection method. Numerical results show that the proposed algorithm gives higher peak SNR, lower temporal error, and lower spatial error than motion-compensated inverse filtering and Lucy-Richardson deconvolution algorithm, which are two state-of-the-art LCD deblurring algorithms.

  17. The solution of the Elrod algorithm for a dynamically loaded journal bearing using multigrid techniques

    Science.gov (United States)

    Woods, C. M.; Brewe, D. E.

    1989-01-01

    A numerical solution to a theoretical model of vapor cavitation in a dynamically loaded journal bearing is developed utilizing a multigrid iteration technique. The method is compared with a noniterative approach in terms of computational time and accuracy. The computational model is based on the Elrod algorithm, a control volume approach to the Reynolds equation which mimics the Jakobsson-Floberg and Olsson cavitation theory. Besides accounting for a moving cavitation boundary and conservation of mass at the boundary, it also conserves mass within the cavitated region via a smeared mass or striated flow extending to both surfaces in the film gap. The mixed nature of the equations (parabolic in the full film zone and hyperbolic in the cavitated zone) coupled with the dynamic aspects of the problem create interesting difficulties for the present solution approach. Emphasis is placed on the methods found to eliminate solution instabilities. Excellent results are obtained for both accuracy and reduction of computational time.

  18. Use of artificial bee colonies algorithm as numerical approximation of differential equations solution

    Science.gov (United States)

    Fikri, Fariz Fahmi; Nuraini, Nuning

    2018-03-01

    The differential equation is one of the branches in mathematics which is closely related to human life problems. Some problems that occur in our life can be modeled into differential equations as well as systems of differential equations such as the Lotka-Volterra model and SIR model. Therefore, solving a problem of differential equations is very important. Some differential equations are difficult to solve, so numerical methods are needed to solve that problems. Some numerical methods for solving differential equations that have been widely used are Euler Method, Heun Method, Runge-Kutta and others. However, some of these methods still have some restrictions that cause the method cannot be used to solve more complex problems such as an evaluation interval that we cannot change freely. New methods are needed to improve that problems. One of the method that can be used is the artificial bees colony algorithm. This algorithm is one of metaheuristic algorithm method, which can come out from local search space and do exploration in solution search space so that will get better solution than other method.

  19. SOFTWARE SOLUTIONS FOR ARDL MODELS

    Directory of Open Access Journals (Sweden)

    Nicolae-Marius JULA

    2015-07-01

    Full Text Available VAR type models can be used only for stationary time series. Causality analyses through econometric models need that series to have the same integrated order. Usually, when constraining the series to comply these restrictions (e.g. by differentiating, economic interpretation of the outcomes may become difficult. Recent solution for mitigating these problems is the use of ARDL (autoregressive distributed lag models. We present implementation in E-Views of these models and we test the impact of exchange rate on consumer price index.

  20. Model Checking Algorithms for CTMDPs

    DEFF Research Database (Denmark)

    Buchholz, Peter; Hahn, Ernst Moritz; Hermanns, Holger

    2011-01-01

    Continuous Stochastic Logic (CSL) can be interpreted over continuoustime Markov decision processes (CTMDPs) to specify quantitative properties of stochastic systems that allow some external control. Model checking CSL formulae over CTMDPs requires then the computation of optimal control strategie...

  1. An efficient algorithm for computation of solitary wave solutions to ...

    Indian Academy of Sciences (India)

    2017-09-08

    Sep 8, 2017 ... Nonlinear mathematical problems and their solutions attain much attention in solitary waves. In soliton theory, an efficient tool to attain various types of soliton solutions is the exp ( − φ ( ζ ) ) -expansion technique. This article is devoted to find exact travelling wave solutions of Drinfeld–Sokolov equation via a ...

  2. An efficient algorithm for computation of solitary wave solutions to ...

    Indian Academy of Sciences (India)

    KAMRAN AYUB

    2017-09-08

    Sep 8, 2017 ... Abstract. Nonlinear mathematical problems and their solutions attain much attention in solitary waves. In soliton theory, an efficient tool to attain various types of soliton solutions is the exp(−ϕ(ζ))-expansion technique. This article is devoted to find exact travelling wave solutions of Drinfeld–Sokolov equation ...

  3. Application of genetic algorithm in radio ecological models parameter determination

    International Nuclear Information System (INIS)

    Pantelic, G.

    2006-01-01

    The method of genetic algorithms was used to determine the biological half-life of 137 Cs in cow milk after the accident in Chernobyl. Methodologically genetic algorithms are based on the fact that natural processes tend to optimize themselves and therefore this method should be more efficient in providing optimal solutions in the modeling of radio ecological and environmental events. The calculated biological half-life of 137 Cs in milk is (32 ± 3) days and transfer coefficient from grass to milk is (0.019 ± 0.005). (authors)

  4. Application of genetic algorithm in radio ecological models parameter determination

    Energy Technology Data Exchange (ETDEWEB)

    Pantelic, G. [Institute of Occupatioanl Health and Radiological Protection ' Dr Dragomir Karajovic' , Belgrade (Serbia)

    2006-07-01

    The method of genetic algorithms was used to determine the biological half-life of 137 Cs in cow milk after the accident in Chernobyl. Methodologically genetic algorithms are based on the fact that natural processes tend to optimize themselves and therefore this method should be more efficient in providing optimal solutions in the modeling of radio ecological and environmental events. The calculated biological half-life of 137 Cs in milk is (32 {+-} 3) days and transfer coefficient from grass to milk is (0.019 {+-} 0.005). (authors)

  5. Rethinking exchange market models as optimization algorithms

    Science.gov (United States)

    Luquini, Evandro; Omar, Nizam

    2018-02-01

    The exchange market model has mainly been used to study the inequality problem. Although the human society inequality problem is very important, the exchange market models dynamics until stationary state and its capability of ranking individuals is interesting in itself. This study considers the hypothesis that the exchange market model could be understood as an optimization procedure. We present herein the implications for algorithmic optimization and also the possibility of a new family of exchange market models

  6. Fuzzy audit risk modeling algorithm

    Directory of Open Access Journals (Sweden)

    Zohreh Hajihaa

    2011-07-01

    Full Text Available Fuzzy logic has created suitable mathematics for making decisions in uncertain environments including professional judgments. One of the situations is to assess auditee risks. During recent years, risk based audit (RBA has been regarded as one of the main tools to fight against fraud. The main issue in RBA is to determine the overall audit risk an auditor accepts, which impact the efficiency of an audit. The primary objective of this research is to redesign the audit risk model (ARM proposed by auditing standards. The proposed model of this paper uses fuzzy inference systems (FIS based on the judgments of audit experts. The implementation of proposed fuzzy technique uses triangular fuzzy numbers to express the inputs and Mamdani method along with center of gravity are incorporated for defuzzification. The proposed model uses three FISs for audit, inherent and control risks, and there are five levels of linguistic variables for outputs. FISs include 25, 25 and 81 rules of if-then respectively and officials of Iranian audit experts confirm all the rules.

  7. Model parameters estimation and sensitivity by genetic algorithms

    International Nuclear Information System (INIS)

    Marseguerra, Marzio; Zio, Enrico; Podofillini, Luca

    2003-01-01

    In this paper we illustrate the possibility of extracting qualitative information on the importance of the parameters of a model in the course of a Genetic Algorithms (GAs) optimization procedure for the estimation of such parameters. The Genetic Algorithms' search of the optimal solution is performed according to procedures that resemble those of natural selection and genetics: an initial population of alternative solutions evolves within the search space through the four fundamental operations of parent selection, crossover, replacement, and mutation. During the search, the algorithm examines a large amount of solution points which possibly carries relevant information on the underlying model characteristics. A possible utilization of this information amounts to create and update an archive with the set of best solutions found at each generation and then to analyze the evolution of the statistics of the archive along the successive generations. From this analysis one can retrieve information regarding the speed of convergence and stabilization of the different control (decision) variables of the optimization problem. In this work we analyze the evolution strategy followed by a GA in its search for the optimal solution with the aim of extracting information on the importance of the control (decision) variables of the optimization with respect to the sensitivity of the objective function. The study refers to a GA search for optimal estimates of the effective parameters in a lumped nuclear reactor model of literature. The supporting observation is that, as most optimization procedures do, the GA search evolves towards convergence in such a way to stabilize first the most important parameters of the model and later those which influence little the model outputs. In this sense, besides estimating efficiently the parameters values, the optimization approach also allows us to provide a qualitative ranking of their importance in contributing to the model output. The

  8. Consistent Algorithms Marching Along Characteristics for the Numerical Solution of the Boltzmann Equation

    Directory of Open Access Journals (Sweden)

    Nilson C. Roberty

    2011-01-01

    Full Text Available We introduce algorithms marching over a polygonal mesh with elements consistent with the propagation directions of the particle (radiation flux. The decision for adopting this kind of mesh to solve the one-speed Boltzmann transport equation is due to characteristics of the domain of the transport operator which controls derivatives only in the direction of propagation of the particles (radiation flux in the absorbing and scattering media. This a priori adaptivity has the advantages that it formulates a consistent scheme which makes appropriate the application of the Lax equivalence theorem framework to the problem. In this work, we present the main functional spaces involved in the formalism and a description of the algorithms for the mesh generation and the transport equation solution. Some numerical examples related to the solution of a transmission problem in a high-contrast model with absorption and scattering are presented. Also, a comparison with benchmarks problems for source and reactor criticality simulations shows the compatibility between calculations with the algorithms proposed here and theoretical results.

  9. General Quantum Meet-in-the-Middle Search Algorithm Based on Target Solution of Fixed Weight

    Science.gov (United States)

    Fu, Xiang-Qun; Bao, Wan-Su; Wang, Xiang; Shi, Jian-Hong

    2016-10-01

    Similar to the classical meet-in-the-middle algorithm, the storage and computation complexity are the key factors that decide the efficiency of the quantum meet-in-the-middle algorithm. Aiming at the target vector of fixed weight, based on the quantum meet-in-the-middle algorithm, the algorithm for searching all n-product vectors with the same weight is presented, whose complexity is better than the exhaustive search algorithm. And the algorithm can reduce the storage complexity of the quantum meet-in-the-middle search algorithm. Then based on the algorithm and the knapsack vector of the Chor-Rivest public-key crypto of fixed weight d, we present a general quantum meet-in-the-middle search algorithm based on the target solution of fixed weight, whose computational complexity is \\sumj = 0d {(O(\\sqrt {Cn - k + 1d - j }) + O(C_kj log C_k^j))} with Σd i =0 Ck i memory cost. And the optimal value of k is given. Compared to the quantum meet-in-the-middle search algorithm for knapsack problem and the quantum algorithm for searching a target solution of fixed weight, the computational complexity of the algorithm is lower. And its storage complexity is smaller than the quantum meet-in-the-middle-algorithm. Supported by the National Basic Research Program of China under Grant No. 2013CB338002 and the National Natural Science Foundation of China under Grant No. 61502526

  10. An efficient algorithm for computation of solitary wave solutions to ...

    Indian Academy of Sciences (India)

    KAMRAN AYUB

    2017-09-08

    Sep 8, 2017 ... exact solitary wave solutions of nonlinear differential equations is a great achievement. In nonlinear physical sciences, an essential contribution is of exact solutions and because of this we can study physical behaviours and discus more features of the problem which give direction to more applications.

  11. Model Checking Algorithms for Markov Reward Models

    NARCIS (Netherlands)

    Cloth, Lucia; Cloth, L.

    2006-01-01

    Model checking Markov reward models unites two different approaches of model-based system validation. On the one hand, Markov reward models have a long tradition in model-based performance and dependability evaluation. On the other hand, a formal method like model checking allows for the precise

  12. Design of attitude solution algorithm for tail-sitter VTOL UAV

    Directory of Open Access Journals (Sweden)

    Donghui LIU

    2016-02-01

    Full Text Available The tail-sitter Vertical Takeoff and Landing (VTOL Unmanned Aerial Vehicle(UAV, flying in a fixed-wing model, overcomes many shortcomings of traditional fixed-wing UAVs, and inherits the advantage of high overall efficiency, which means it has great development potential and very broad application prospects. The attitude of tail-sitter VTOL UAV shows a wide change range in its takeoff and landing stages, and when the attitude sensor changes more than 90 degrees in pitch direction, the Euler angles converted by the Quaternions will have singular points, which means gimbal deadlock appears. From the solution algorithm, this paper provides a method of changing the order of rotation to avoid the appearance of singular points. The results show that this method can be well applied to the attitude solution of the VTOL UAV.

  13. Worm algorithm for the CPN−1 model

    Directory of Open Access Journals (Sweden)

    Tobias Rindlisbacher

    2017-05-01

    Full Text Available The CPN−1 model in 2D is an interesting toy model for 4D QCD as it possesses confinement, asymptotic freedom and a non-trivial vacuum structure. Due to the lower dimensionality and the absence of fermions, the computational cost for simulating 2D CPN−1 on the lattice is much lower than that for simulating 4D QCD. However, to our knowledge, no efficient algorithm for simulating the lattice CPN−1 model for N>2 has been tested so far, which also works at finite density. To this end we propose a new type of worm algorithm which is appropriate to simulate the lattice CPN−1 model in a dual, flux-variables based representation, in which the introduction of a chemical potential does not give rise to any complications. In addition to the usual worm moves where a defect is just moved from one lattice site to the next, our algorithm additionally allows for worm-type moves in the internal variable space of single links, which accelerates the Monte Carlo evolution. We use our algorithm to compare the two popular CPN−1 lattice actions and exhibit marked differences in their approach to the continuum limit.

  14. Algorithms and Models for the Web Graph

    NARCIS (Netherlands)

    Gleich, David F.; Komjathy, Julia; Litvak, Nelli

    2015-01-01

    This volume contains the papers presented at WAW2015, the 12th Workshop on Algorithms and Models for the Web-Graph held during December 10–11, 2015, in Eindhoven. There were 24 submissions. Each submission was reviewed by at least one, and on average two, Program Committee members. The committee

  15. Model based development of engine control algorithms

    NARCIS (Netherlands)

    Dekker, H.J.; Sturm, W.L.

    1996-01-01

    Model based development of engine control systems has several advantages. The development time and costs are strongly reduced because much of the development and optimization work is carried out by simulating both engine and control system. After optimizing the control algorithm it can be executed

  16. Optimization in engineering models and algorithms

    CERN Document Server

    Sioshansi, Ramteen

    2017-01-01

    This textbook covers the fundamentals of optimization, including linear, mixed-integer linear, nonlinear, and dynamic optimization techniques, with a clear engineering focus. It carefully describes classical optimization models and algorithms using an engineering problem-solving perspective, and emphasizes modeling issues using many real-world examples related to a variety of application areas. Providing an appropriate blend of practical applications and optimization theory makes the text useful to both practitioners and students, and gives the reader a good sense of the power of optimization and the potential difficulties in applying optimization to modeling real-world systems. The book is intended for undergraduate and graduate-level teaching in industrial engineering and other engineering specialties. It is also of use to industry practitioners, due to the inclusion of real-world applications, opening the door to advanced courses on both modeling and algorithm development within the industrial engineering ...

  17. IIR Filter Modeling Using an Algorithm Inspired on Electromagnetism

    Directory of Open Access Journals (Sweden)

    Cuevas-Jiménez E.

    2013-01-01

    Full Text Available Infinite-impulse-response (IIR filtering provides a powerful approach for solving a variety of problems. However, its design represents a very complicated task, since the error surface of IIR filters is generally multimodal, global optimization techniques are required in order to avoid local minima. In this paper, a new method based on the Electromagnetism-Like Optimization Algorithm (EMO is proposed for IIR filter modeling. EMO originates from the electro-magnetism theory of physics by assuming potential solutions as electrically charged particles which spread around the solution space. The charge of each particle depends on its objective function value. This algorithm employs a collective attraction-repulsion mechanism to move the particles towards optimality. The experimental results confirm the high performance of the proposed method in solving various benchmark identification problems.

  18. An Enhanced Artificial Bee Colony Algorithm with Solution Acceptance Rule and Probabilistic Multisearch

    Directory of Open Access Journals (Sweden)

    Alkın Yurtkuran

    2016-01-01

    Full Text Available The artificial bee colony (ABC algorithm is a popular swarm based technique, which is inspired from the intelligent foraging behavior of honeybee swarms. This paper proposes a new variant of ABC algorithm, namely, enhanced ABC with solution acceptance rule and probabilistic multisearch (ABC-SA to address global optimization problems. A new solution acceptance rule is proposed where, instead of greedy selection between old solution and new candidate solution, worse candidate solutions have a probability to be accepted. Additionally, the acceptance probability of worse candidates is nonlinearly decreased throughout the search process adaptively. Moreover, in order to improve the performance of the ABC and balance the intensification and diversification, a probabilistic multisearch strategy is presented. Three different search equations with distinctive characters are employed using predetermined search probabilities. By implementing a new solution acceptance rule and a probabilistic multisearch approach, the intensification and diversification performance of the ABC algorithm is improved. The proposed algorithm has been tested on well-known benchmark functions of varying dimensions by comparing against novel ABC variants, as well as several recent state-of-the-art algorithms. Computational results show that the proposed ABC-SA outperforms other ABC variants and is superior to state-of-the-art algorithms proposed in the literature.

  19. Asymmetry in some common assignment algorithms: the dispersion factor solution

    OpenAIRE

    T de la Barra; B Pérez

    1986-01-01

    Many common assignment algorithms are based on Dial's original design to determine the paths that trip makers will follow from a given origin to destination centroids. The purpose of this paper is to show that the rules that have to be applied result in two unwanted properties. The first is that trips assigned from an origin centroid i to a destination j can be dramatically different to those resulting from centroid j to centroid i , even if the number of trips is the same and the network is ...

  20. Estimating the solute transport parameters of the spatial fractional advection-dispersion equation using Bees Algorithm

    Science.gov (United States)

    Mehdinejadiani, Behrouz

    2017-08-01

    This study represents the first attempt to estimate the solute transport parameters of the spatial fractional advection-dispersion equation using Bees Algorithm. The numerical studies as well as the experimental studies were performed to certify the integrity of Bees Algorithm. The experimental ones were conducted in a sandbox for homogeneous and heterogeneous soils. A detailed comparative study was carried out between the results obtained from Bees Algorithm and those from Genetic Algorithm and LSQNONLIN routines in FracFit toolbox. The results indicated that, in general, the Bees Algorithm much more accurately appraised the sFADE parameters in comparison with Genetic Algorithm and LSQNONLIN, especially in the heterogeneous soil and for α values near to 1 in the numerical study. Also, the results obtained from Bees Algorithm were more reliable than those from Genetic Algorithm. The Bees Algorithm showed the relative similar performances for all cases, while the Genetic Algorithm and the LSQNONLIN yielded different performances for various cases. The performance of LSQNONLIN strongly depends on the initial guess values so that, compared to the Genetic Algorithm, it can more accurately estimate the sFADE parameters by taking into consideration the suitable initial guess values. To sum up, the Bees Algorithm was found to be very simple, robust and accurate approach to estimate the transport parameters of the spatial fractional advection-dispersion equation.

  1. Markov chains models, algorithms and applications

    CERN Document Server

    Ching, Wai-Ki; Ng, Michael K; Siu, Tak-Kuen

    2013-01-01

    This new edition of Markov Chains: Models, Algorithms and Applications has been completely reformatted as a text, complete with end-of-chapter exercises, a new focus on management science, new applications of the models, and new examples with applications in financial risk management and modeling of financial data.This book consists of eight chapters.  Chapter 1 gives a brief introduction to the classical theory on both discrete and continuous time Markov chains. The relationship between Markov chains of finite states and matrix theory will also be highlighted. Some classical iterative methods

  2. Modelling Evolutionary Algorithms with Stochastic Differential Equations.

    Science.gov (United States)

    Heredia, Jorge Pérez

    2017-11-20

    There has been renewed interest in modelling the behaviour of evolutionary algorithms (EAs) by more traditional mathematical objects, such as ordinary differential equations or Markov chains. The advantage is that the analysis becomes greatly facilitated due to the existence of well established methods. However, this typically comes at the cost of disregarding information about the process. Here, we introduce the use of stochastic differential equations (SDEs) for the study of EAs. SDEs can produce simple analytical results for the dynamics of stochastic processes, unlike Markov chains which can produce rigorous but unwieldy expressions about the dynamics. On the other hand, unlike ordinary differential equations (ODEs), they do not discard information about the stochasticity of the process. We show that these are especially suitable for the analysis of fixed budget scenarios and present analogues of the additive and multiplicative drift theorems from runtime analysis. In addition, we derive a new more general multiplicative drift theorem that also covers non-elitist EAs. This theorem simultaneously allows for positive and negative results, providing information on the algorithm's progress even when the problem cannot be optimised efficiently. Finally, we provide results for some well-known heuristics namely Random Walk (RW), Random Local Search (RLS), the (1+1) EA, the Metropolis Algorithm (MA), and the Strong Selection Weak Mutation (SSWM) algorithm.

  3. On randomized algorithms for numerical solution of applied Fredholm integral equations of the second kind

    Science.gov (United States)

    Voytishek, Anton V.; Shipilov, Nikolay M.

    2017-11-01

    In this paper, the systematization of numerical (implemented on a computer) randomized functional algorithms for approximation of a solution of Fredholm integral equation of the second kind is carried out. Wherein, three types of such algorithms are distinguished: the projection, the mesh and the projection-mesh methods. The possibilities for usage of these algorithms for solution of practically important problems is investigated in detail. The disadvantages of the mesh algorithms, related to the necessity of calculation values of the kernels of integral equations in fixed points, are identified. On practice, these kernels have integrated singularities, and calculation of their values is impossible. Thus, for applied problems, related to solving Fredholm integral equation of the second kind, it is expedient to use not mesh, but the projection and the projection-mesh randomized algorithms.

  4. An Efficient Algorithm for Partitioning and Authenticating Problem-Solutions of eLeaming Contents

    Science.gov (United States)

    Dewan, Jahangir; Chowdhury, Morshed; Batten, Lynn

    2013-01-01

    Content authenticity and correctness is one of the important challenges in eLearning as there can be many solutions to one specific problem in cyber space. Therefore, the authors feel it is necessary to map problems to solutions using graph partition and weighted bipartite matching. This article proposes an efficient algorithm to partition…

  5. Sparse modeling theory, algorithms, and applications

    CERN Document Server

    Rish, Irina

    2014-01-01

    ""A comprehensive, clear, and well-articulated book on sparse modeling. This book will stand as a prime reference to the research community for many years to come.""-Ricardo Vilalta, Department of Computer Science, University of Houston""This book provides a modern introduction to sparse methods for machine learning and signal processing, with a comprehensive treatment of both theory and algorithms. Sparse Modeling is an ideal book for a first-year graduate course.""-Francis Bach, INRIA - École Normale Supřieure, Paris

  6. Performance comparison of genetic algorithms and particle swarm optimization for model integer programming bus timetabling problem

    Science.gov (United States)

    Wihartiko, F. D.; Wijayanti, H.; Virgantari, F.

    2018-03-01

    Genetic Algorithm (GA) is a common algorithm used to solve optimization problems with artificial intelligence approach. Similarly, the Particle Swarm Optimization (PSO) algorithm. Both algorithms have different advantages and disadvantages when applied to the case of optimization of the Model Integer Programming for Bus Timetabling Problem (MIPBTP), where in the case of MIPBTP will be found the optimal number of trips confronted with various constraints. The comparison results show that the PSO algorithm is superior in terms of complexity, accuracy, iteration and program simplicity in finding the optimal solution.

  7. Genetic Algorithm Optimization of Artificial Neural Networks for Hydrological Modelling

    Science.gov (United States)

    Abrahart, R. J.

    2004-05-01

    This paper will consider the case for genetic algorithm optimization in the development of an artificial neural network model. It will provide a methodological evaluation of reported investigations with respect to hydrological forecasting and prediction. The intention in such operations is to develop a superior modelling solution that will be: \\begin{itemize} more accurate in terms of output precision and model estimation skill; more tractable in terms of personal requirements and end-user control; and/or more robust in terms of conceptual and mechanical power with respect to adverse conditions. The genetic algorithm optimization toolbox could be used to perform a number of specific roles or purposes and it is the harmonious and supportive relationship between neural networks and genetic algorithms that will be highlighted and assessed. There are several neural network mechanisms and procedures that could be enhanced and potential benefits are possible at different stages in the design and construction of an operational hydrological model e.g. division of inputs; identification of structure; initialization of connection weights; calibration of connection weights; breeding operations between successful models; and output fusion associated with the development of ensemble solutions. Each set of opportunities will be discussed and evaluated. Two strategic questions will also be considered: [i] should optimization be conducted as a set of small individual procedures or as one large holistic operation; [ii] what specific function or set of weighted vectors should be optimized in a complex software product e.g. timings, volumes, or quintessential hydrological attributes related to the 'problem situation' - that might require the development flood forecasting, drought estimation, or record infilling applications. The paper will conclude with a consideration of hydrological forecasting solutions developed on the combined methodologies of co-operative co-evolution and

  8. Water evaporation algorithm: A new metaheuristic algorithm towards the solution of optimal power flow

    Directory of Open Access Journals (Sweden)

    Anulekha Saha

    2017-12-01

    Full Text Available A relatively new technique to solve the optimal power flow (OPF problem inspired by the evaporation (vaporization of small quantity water particles from dense surfaces is presented in this paper. IEEE 30 bus and IEEE 118 bus test systems are assessed for various objectives to determine water evaporation algorithm’s (WEA efficiency in handling the OPF problem after satisfying constraints. Comparative study with other established techniques demonstrate competitiveness of WEA in treating varied objectives. It achieved superior results for all the objectives considered. The algorithm is found to minimize its objective values by great margins even in case of large test system. Statistical analysis of all the cases using Wilcoxon’s signed rank test resulted in p-values much lower than the required value of 0.05, thereby establishing the robustness of the applied technique. Best performance of the algorithm are obtained for voltage deviation minimization and voltage stability index minimization objectives in case of IEEE 30 and IEEE 118 bus test systems respectively.

  9. Link mining models, algorithms, and applications

    CERN Document Server

    Yu, Philip S; Faloutsos, Christos

    2010-01-01

    This book presents in-depth surveys and systematic discussions on models, algorithms and applications for link mining. Link mining is an important field of data mining. Traditional data mining focuses on 'flat' data in which each data object is represented as a fixed-length attribute vector. However, many real-world data sets are much richer in structure, involving objects of multiple types that are related to each other. Hence, recently link mining has become an emerging field of data mining, which has a high impact in various important applications such as text mining, social network analysi

  10. Genetic Algorithms Principles Towards Hidden Markov Model

    Directory of Open Access Journals (Sweden)

    Nabil M. Hewahi

    2011-10-01

    Full Text Available In this paper we propose a general approach based on Genetic Algorithms (GAs to evolve Hidden Markov Models (HMM. The problem appears when experts assign probability values for HMM, they use only some limited inputs. The assigned probability values might not be accurate to serve in other cases related to the same domain. We introduce an approach based on GAs to find
    out the suitable probability values for the HMM to be mostly correct in more cases than what have been used to assign the probability values.

  11. Routing Flow-Shop with Buffers and Ready Times – Comparison of Selected Solution Algorithms

    OpenAIRE

    Józefczyk Jerzy; Markowski Michał; Balgabaeva Ljazat

    2014-01-01

    This article extends the former results concerning the routing flow-shop problem to minimize the makespan on the case with buffers, non-zero ready times and different speeds of machines. The corresponding combinatorial optimization problem is formulated. The exact as well as four heuristic solution algorithms are presented. The branch and bound approach is applied for the former one. The heuristic algorithms employ known constructive idea proposed for the former version of the problem as well...

  12. Optimization Solutions for Improving the Performance of the Parallel Reduction Algorithm Using Graphics Processing Units

    Directory of Open Access Journals (Sweden)

    Ion LUNGU

    2012-01-01

    Full Text Available In this paper, we research, analyze and develop optimization solutions for the parallel reduction function using graphics processing units (GPUs that implement the Compute Unified Device Architecture (CUDA, a modern and novel approach for improving the software performance of data processing applications and algorithms. Many of these applications and algorithms make use of the reduction function in their computational steps. After having designed the function and its algorithmic steps in CUDA, we have progressively developed and implemented optimization solutions for the reduction function. In order to confirm, test and evaluate the solutions' efficiency, we have developed a custom tailored benchmark suite. We have analyzed the obtained experimental results regarding: the comparison of the execution time and bandwidth when using graphic processing units covering the main CUDA architectures (Tesla GT200, Fermi GF100, Kepler GK104 and a central processing unit; the data type influence; the binary operator's influence.

  13. SPECIAL LIBRARIES OF FRAGMENTS OF ALGORITHMIC NETWORKS TO AUTOMATE THE DEVELOPMENT OF ALGORITHMIC MODELS

    Directory of Open Access Journals (Sweden)

    V. E. Marley

    2015-01-01

    Full Text Available Summary. The concept of algorithmic models appeared from the algorithmic approach in which the simulated object, the phenomenon appears in the form of process, subject to strict rules of the algorithm, which placed the process of operation of the facility. Under the algorithmic model is the formalized description of the scenario subject specialist for the simulated process, the structure of which is comparable with the structure of the causal and temporal relationships between events of the process being modeled, together with all information necessary for its software implementation. To represent the structure of algorithmic models used algorithmic network. Normally, they were defined as loaded finite directed graph, the vertices which are mapped to operators and arcs are variables, bound by operators. The language of algorithmic networks has great features, the algorithms that it can display indifference the class of all random algorithms. In existing systems, automation modeling based on algorithmic nets, mainly used by operators working with real numbers. Although this reduces their ability, but enough for modeling a wide class of problems related to economy, environment, transport, technical processes. The task of modeling the execution of schedules and network diagrams is relevant and useful. There are many counting systems, network graphs, however, the monitoring process based analysis of gaps and terms of graphs, no analysis of prediction execution schedule or schedules. The library is designed to build similar predictive models. Specifying source data to obtain a set of projections from which to choose one and take it for a new plan.

  14. Numerical solution of dynamic equilibrium models under Poisson uncertainty

    DEFF Research Database (Denmark)

    Posch, Olaf; Trimborn, Timo

    2013-01-01

    We propose a simple and powerful numerical algorithm to compute the transition process in continuous-time dynamic equilibrium models with rare events. In this paper we transform the dynamic system of stochastic differential equations into a system of functional differential equations...... of the retarded type. We apply the Waveform Relaxation algorithm, i.e., we provide a guess of the policy function and solve the resulting system of (deterministic) ordinary differential equations by standard techniques. For parametric restrictions, analytical solutions to the stochastic growth model and a novel...... solution to Lucas' endogenous growth model under Poisson uncertainty are used to compute the exact numerical error. We show how (potential) catastrophic events such as rare natural disasters substantially affect the economic decisions of households....

  15. An efficient parallel algorithm for the solution of a tridiagonal linear system of equations

    Science.gov (United States)

    Stone, H. S.

    1971-01-01

    Tridiagonal linear systems of equations are solved on conventional serial machines in a time proportional to N, where N is the number of equations. The conventional algorithms do not lend themselves directly to parallel computations on computers of the ILLIAC IV class, in the sense that they appear to be inherently serial. An efficient parallel algorithm is presented in which computation time grows as log sub 2 N. The algorithm is based on recursive doubling solutions of linear recurrence relations, and can be used to solve recurrence relations of all orders.

  16. Models and Algorithms for Tracking Target with Coordinated Turn Motion

    Directory of Open Access Journals (Sweden)

    Xianghui Yuan

    2014-01-01

    Full Text Available Tracking target with coordinated turn (CT motion is highly dependent on the models and algorithms. First, the widely used models are compared in this paper—coordinated turn (CT model with known turn rate, augmented coordinated turn (ACT model with Cartesian velocity, ACT model with polar velocity, CT model using a kinematic constraint, and maneuver centered circular motion model. Then, in the single model tracking framework, the tracking algorithms for the last four models are compared and the suggestions on the choice of models for different practical target tracking problems are given. Finally, in the multiple models (MM framework, the algorithm based on expectation maximization (EM algorithm is derived, including both the batch form and the recursive form. Compared with the widely used interacting multiple model (IMM algorithm, the EM algorithm shows its effectiveness.

  17. Ischemic postconditioning: experimental models and protocol algorithms.

    Science.gov (United States)

    Skyschally, Andreas; van Caster, Patrick; Iliodromitis, Efstathios K; Schulz, Rainer; Kremastinos, Dimitrios T; Heusch, Gerd

    2009-09-01

    Ischemic postconditioning, a simple mechanical maneuver at the onset of reperfusion, reduces infarct size after ischemia/reperfusion. After its first description in 2003 by Zhao et al. numerous experimental studies have investigated this protective phenomenon. Whereas the underlying mechanisms and signal transduction are not yet understood in detail, infarct size reduction by ischemic postconditioning was confirmed in all species tested so far, including man. We have now reviewed the literature with focus on experimental models and protocols to better understand the determinants of protection by ischemic postconditioning or lack of it. Only studies with infarct size as unequivocal endpoint were considered. In all species and models, the duration of index ischemia and the protective protocol algorithm impact on the outcome of ischemic postconditioning, and gender, age, and myocardial temperature contribute.

  18. Nonlinear model predictive control theory and algorithms

    CERN Document Server

    Grüne, Lars

    2017-01-01

    This book offers readers a thorough and rigorous introduction to nonlinear model predictive control (NMPC) for discrete-time and sampled-data systems. NMPC schemes with and without stabilizing terminal constraints are detailed, and intuitive examples illustrate the performance of different NMPC variants. NMPC is interpreted as an approximation of infinite-horizon optimal control so that important properties like closed-loop stability, inverse optimality and suboptimality can be derived in a uniform manner. These results are complemented by discussions of feasibility and robustness. An introduction to nonlinear optimal control algorithms yields essential insights into how the nonlinear optimization routine—the core of any nonlinear model predictive controller—works. Accompanying software in MATLAB® and C++ (downloadable from extras.springer.com/), together with an explanatory appendix in the book itself, enables readers to perform computer experiments exploring the possibilities and limitations of NMPC. T...

  19. A genetic algorithm solution for combinatorial problems - the nuclear core reload example

    Energy Technology Data Exchange (ETDEWEB)

    Schirru, R.; Silva, F.C. [Universidade Federal, Rio de Janeiro, RJ (Brazil). Coordenacao dos Programas de Pos-graduacao de Engenharia; Pereira, C.M.N.A. [Instituto de Engenharia Nuclear (IEN), Rio de Janeiro, RJ (Brazil); Chapot, J.L.C. [FURNAS, Rio de Janeiro, RJ (Brazil)

    1997-12-01

    This paper presents a solution to Traveling Salesman Problem based upon genetic algorithms (GA), using the classic crossover, but avoiding the feasibility problem in offspring individuals, allowing the natural evolution of the GA without introduction of heuristics in the genetic crossover operator. The genetic model presented, that we call the List Model (LM) is based on the encoding and decoding genotype in the way to always generate a phenotype that has a valid structure, over which will be applied the fitness, represented by the total distance. The main purpose of this work was to develop the basis for a new genetic model to be used in the reload of nuclear core of a PWR. In a generic way, this problem can be interpreted as a a search of the optimal combination of N different fuel elements in N nuclear core `holes`, where each combination or load pattern, determines the neutron flux shape and its associate peak factor. The goal is to find out the load pattern that minimizes the peak factor and consequently maximize the useful life of the nuclear fuel. The GA with the List Model was applied to the Angra-1 PWR reload problem and the results are remarkably better than the ones used in the last fuel cycle. (author). 12 refs., 3 figs., 2 tabs.

  20. Development of an inter-layer solute transport algorithm for SOLTR computer program. Part 1. The algorithm

    International Nuclear Information System (INIS)

    Miller, I.; Roman, K.

    1979-12-01

    In order to perform studies of the influence of regional groundwater flow systems on the long-term performance of potential high-level nuclear waste repositories, it was determined that an adequate computer model would have to consider the full three-dimensional flow system. Golder Associates' SOLTR code, while three-dimensional, has an overly simple algorithm for simulating the passage of radionuclides from one aquifier to another above or below it. Part 1 of this report describes the algorithm developed to provide SOLTR with an improved capability for simulating interaquifer transport

  1. Verification of fluid-structure-interaction algorithms through the method of manufactured solutions for actuator-line applications

    Science.gov (United States)

    Vijayakumar, Ganesh; Sprague, Michael

    2017-11-01

    Demonstrating expected convergence rates with spatial- and temporal-grid refinement is the ``gold standard'' of code and algorithm verification. However, the lack of analytical solutions and generating manufactured solutions presents challenges for verifying codes for complex systems. The application of the method of manufactured solutions (MMS) for verification for coupled multi-physics phenomena like fluid-structure interaction (FSI) has only seen recent investigation. While many FSI algorithms for aeroelastic phenomena have focused on boundary-resolved CFD simulations, the actuator-line representation of the structure is widely used for FSI simulations in wind-energy research. In this work, we demonstrate the verification of an FSI algorithm using MMS for actuator-line CFD simulations with a simplified structural model. We use a manufactured solution for the fluid velocity field and the displacement of the SMD system. We demonstrate the convergence of both the fluid and structural solver to second-order accuracy with grid and time-step refinement. This work was funded by the U.S. Department of Energy, Office of Energy Efficiency and Renewable Energy, Wind Energy Technologies Office, under Contract No. DE-AC36-08-GO28308 with the National Renewable Energy Laboratory.

  2. Adaptive numerical algorithms in space weather modeling

    Science.gov (United States)

    Tóth, Gábor; van der Holst, Bart; Sokolov, Igor V.; De Zeeuw, Darren L.; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Najib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav

    2012-02-01

    Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different relevant physics in different domains. A multi-physics system can be modeled by a software framework comprising several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solarwind Roe-type Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamic (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit

  3. Adaptive numerical algorithms in space weather modeling

    International Nuclear Information System (INIS)

    Tóth, Gábor; Holst, Bart van der; Sokolov, Igor V.; De Zeeuw, Darren L.; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng Xing; Najib, Dalal; Powell, Kenneth G.; Stout, Quentin F.; Glocer, Alex; Ma, Ying-Juan; Opher, Merav

    2012-01-01

    Space weather describes the various processes in the Sun–Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different relevant physics in different domains. A multi-physics system can be modeled by a software framework comprising several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solarwind Roe-type Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamic (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit

  4. Adaptive Numerical Algorithms in Space Weather Modeling

    Science.gov (United States)

    Toth, Gabor; vanderHolst, Bart; Sokolov, Igor V.; DeZeeuw, Darren; Gombosi, Tamas I.; Fang, Fang; Manchester, Ward B.; Meng, Xing; Nakib, Dalal; Powell, Kenneth G.; hide

    2010-01-01

    Space weather describes the various processes in the Sun-Earth system that present danger to human health and technology. The goal of space weather forecasting is to provide an opportunity to mitigate these negative effects. Physics-based space weather modeling is characterized by disparate temporal and spatial scales as well as by different physics in different domains. A multi-physics system can be modeled by a software framework comprising of several components. Each component corresponds to a physics domain, and each component is represented by one or more numerical models. The publicly available Space Weather Modeling Framework (SWMF) can execute and couple together several components distributed over a parallel machine in a flexible and efficient manner. The framework also allows resolving disparate spatial and temporal scales with independent spatial and temporal discretizations in the various models. Several of the computationally most expensive domains of the framework are modeled by the Block-Adaptive Tree Solar wind Roe Upwind Scheme (BATS-R-US) code that can solve various forms of the magnetohydrodynamics (MHD) equations, including Hall, semi-relativistic, multi-species and multi-fluid MHD, anisotropic pressure, radiative transport and heat conduction. Modeling disparate scales within BATS-R-US is achieved by a block-adaptive mesh both in Cartesian and generalized coordinates. Most recently we have created a new core for BATS-R-US: the Block-Adaptive Tree Library (BATL) that provides a general toolkit for creating, load balancing and message passing in a 1, 2 or 3 dimensional block-adaptive grid. We describe the algorithms of BATL and demonstrate its efficiency and scaling properties for various problems. BATS-R-US uses several time-integration schemes to address multiple time-scales: explicit time stepping with fixed or local time steps, partially steady-state evolution, point-implicit, semi-implicit, explicit/implicit, and fully implicit numerical

  5. Engineering of Algorithms for Hidden Markov models and Tree Distances

    DEFF Research Database (Denmark)

    Sand, Andreas

    speed up all the classical algorithms for analyses and training of hidden Markov models. And I show how two particularly important algorithms, the forward algorithm and the Viterbi algorithm, can be accelerated through a reformulation of the algorithms and a somewhat more complicated parallelization....... Lastly, I show how hidden Markov models can be trained orders of magnitude faster on a given input by rethinking the forward algorithm such that it can automatically adapt itself to the input. Together, these optimization have enabled us to perform analysis of full genomes in a few minutes and thereby...

  6. Genetic Algorithm Approaches to Prebiobiotic Chemistry Modeling

    Science.gov (United States)

    Lohn, Jason; Colombano, Silvano

    1997-01-01

    We model an artificial chemistry comprised of interacting polymers by specifying two initial conditions: a distribution of polymers and a fixed set of reversible catalytic reactions. A genetic algorithm is used to find a set of reactions that exhibit a desired dynamical behavior. Such a technique is useful because it allows an investigator to determine whether a specific pattern of dynamics can be produced, and if it can, the reaction network found can be then analyzed. We present our results in the context of studying simplified chemical dynamics in theorized protocells - hypothesized precursors of the first living organisms. Our results show that given a small sample of plausible protocell reaction dynamics, catalytic reaction sets can be found. We present cases where this is not possible and also analyze the evolved reaction sets.

  7. Parameter Optimization of Single-Diode Model of Photovoltaic Cell Using Memetic Algorithm

    Directory of Open Access Journals (Sweden)

    Yourim Yoon

    2015-01-01

    Full Text Available This study proposes a memetic approach for optimally determining the parameter values of single-diode-equivalent solar cell model. The memetic algorithm, which combines metaheuristic and gradient-based techniques, has the merit of good performance in both global and local searches. First, 10 single algorithms were considered including genetic algorithm, simulated annealing, particle swarm optimization, harmony search, differential evolution, cuckoo search, least squares method, and pattern search; then their final solutions were used as initial vectors for generalized reduced gradient technique. From this memetic approach, we could further improve the accuracy of the estimated solar cell parameters when compared with single algorithm approaches.

  8. Investigation of the Solution Space of Marine Controlled-Source Electromagnetic Inversion Problems By Using a Genetic Algorithm

    Science.gov (United States)

    Hunziker, J.; Thorbecke, J.; Slob, E. C.

    2014-12-01

    Commonly, electromagnetic measurements for exploring and monitoring hydrocarbon reservoirs are inverted for the subsurface conductivity distribution by minimizing the difference between the actual data and a forward modeled dataset. The convergence of the inversion process to the correct solution strongly depends on the shape of the solution space. Since this is a non-linear problem, there exist a multitude of minima of which only the global one provides the correct conductivity values. To easily find the global minimum we desire it to have a broad cone of attraction, while it should also feature a very narrow bottom in order to obtain the subsurface conductivity with high resolution. In this study, we aim to determine which combination of input data corresponds to a favorable shape of the solution space. Since the solution space is N-dimensional, with N being the number of unknown subsurface parameters, plotting it is out of the question. In our approach, we use a genetic algorithm (Goldberg, 1989) to probe the solution space. Such algorithms have the advantage that every run of the same problem will end up at a different solution. Most of these solutions are expected to lie close to the global minimum. A situation where only few runs end up in the global minimum indicates that the solution space consists of a lot of local minima or that the cone of attraction of the global minimum is small. If a lot of runs end up with a similar data-misfit but with a large spread of the subsurface medium parameters in one or more direction, it can be concluded that the chosen data-input is not sensitive with respect to that direction. Compared to the study of Hunziker et al. 2014, we allow also to invert for subsurface boundaries and include more combinations of input datasets. The results so far suggest that it is essential to include the magnetic field in the inversion process in order to find the anisotropic conductivity values. ReferencesGoldberg, D. E., 1989. Genetic

  9. Parametrisation of a Maxwell model for transient tyre forces by means of an extended firefly algorithm

    Directory of Open Access Journals (Sweden)

    Andreas Hackl

    2016-12-01

    Full Text Available Developing functions for advanced driver assistance systems requires very accurate tyre models, especially for the simulation of transient conditions. In the past, parametrisation of a given tyre model based on measurement data showed shortcomings, and the globally optimal solution obtained did not appear to be plausible. In this article, an optimisation strategy is presented, which is able to find plausible and physically feasible solutions by detecting many local outcomes. The firefly algorithm mimics the natural behaviour of fireflies, which use a kind of flashing light to communicate with other members. An algorithm simulating the intensity of the light of a single firefly, diminishing with increasing distances, is implicitly able to detect local solutions on its way to the best solution in the search space. This implicit clustering feature is stressed by an additional explicit clustering step, where local solutions are stored and terminally processed to obtain a large number of possible solutions. The enhanced firefly algorithm will be first applied to the well-known Rastrigin functions and then to the tyre parametrisation problem. It is shown that the firefly algorithm is qualified to find a high number of optimisation solutions, which is required for plausible parametrisation for the given tyre model.

  10. An MPCC Formulation and Its Smooth Solution Algorithm for Continuous Network Design Problem

    Directory of Open Access Journals (Sweden)

    Guangmin Wang

    2017-12-01

    Full Text Available Continuous network design problem (CNDP is searching for a transportation network configuration to minimize the sum of the total system travel time and the investment cost of link capacity expansions by considering that the travellers follow a traditional Wardrop user equilibrium (UE to choose their routes. In this paper, the CNDP model can be formulated as mathematical programs with complementarity constraints (MPCC by describing UE as a non-linear complementarity problem (NCP. To address the difficulty resulting from complementarity constraints in MPCC, they are substituted by the Fischer-Burmeister (FB function, which can be smoothed by the introduction of the smoothing parameter. Therefore, the MPCC can be transformed into a well-behaved non-linear program (NLP by replacing the complementarity constraints with a smooth equation. Consequently, the solver such as LINDOGLOBAL in GAMS can be used to solve the smooth approximate NLP to obtain the solution to MPCC for modelling CNDP. The numerical experiments on the example from the literature demonstrate that the proposed algorithm is feasible.

  11. Integer programming model for optimizing bus timetable using genetic algorithm

    Science.gov (United States)

    Wihartiko, F. D.; Buono, A.; Silalahi, B. P.

    2017-01-01

    Bus timetable gave an information for passengers to ensure the availability of bus services. Timetable optimal condition happened when bus trips frequency could adapt and suit with passenger demand. In the peak time, the number of bus trips would be larger than the off-peak time. If the number of bus trips were more frequent than the optimal condition, it would make a high operating cost for bus operator. Conversely, if the number of trip was less than optimal condition, it would make a bad quality service for passengers. In this paper, the bus timetabling problem would be solved by integer programming model with modified genetic algorithm. Modification was placed in the chromosomes design, initial population recovery technique, chromosomes reconstruction and chromosomes extermination on specific generation. The result of this model gave the optimal solution with accuracy 99.1%.

  12. A deterministic annealing algorithm for approximating a solution of the linearly constrained nonconvex quadratic minimization problem.

    Science.gov (United States)

    Dang, Chuangyin; Liang, Jianqing; Yang, Yang

    2013-03-01

    A deterministic annealing algorithm is proposed for approximating a solution of the linearly constrained nonconvex quadratic minimization problem. The algorithm is derived from applications of a Hopfield-type barrier function in dealing with box constraints and Lagrange multipliers in handling linear equality constraints, and attempts to obtain a solution of good quality by generating a minimum point of a barrier problem for a sequence of descending values of the barrier parameter. For any given value of the barrier parameter, the algorithm searches for a minimum point of the barrier problem in a feasible descent direction, which has a desired property that the box constraints are always satisfied automatically if the step length is a number between zero and one. At each iteration, the feasible descent direction is found by updating Lagrange multipliers with a globally convergent iterative procedure. For any given value of the barrier parameter, the algorithm converges to a stationary point of the barrier problem. Preliminary numerical results show that the algorithm seems effective and efficient. Copyright © 2012 Elsevier Ltd. All rights reserved.

  13. Algorithm for Solution of Direct Kinematic Problem of Multi-sectional Manipulator with Parallel Structure

    Directory of Open Access Journals (Sweden)

    A. L. Lapikov

    2014-01-01

    Full Text Available The article is aimed at creating techniques to study multi-sectional manipulators with parallel structure. To solve this task the analysis in the field concerned was carried out to reveal both advantages and drawbacks of such executive mechanisms and main problems to be encountered in the course of research. The work shows that it is inefficient to create complete mathematical models of multisectional manipulators, which in the context of solving a direct kinematic problem are to derive a functional dependence of location and orientation of the end effector on all the generalized coordinates of the mechanism. The structure of multisectional manipulators was considered, where the sections are platform manipulators of parallel kinematics with six degrees of freedom. The paper offers an algorithm to define location and orientation of the end effector of the manipulator by means of iterative solution of analytical equation of the moving platform plane for each section. The equation for the unknown plane is derived using three points, which are attachment points of the moving platform joints. To define the values of joint coordinates a system of nine non-linear equations is completed. It is necessary to mention that for completion of the equation system are used the equations with the same type of non-linearity. The physical sense of all nine equations of the system is Euclidean distance between the points of the manipulator. The result of algorithm execution is a matrix of homogenous transformation for each section. The correlations describing transformations between adjoining sections of the manipulator are given. An example of the mechanism consisting of three sections is examined. The comparison of theoretical calculations with results obtained on a 3D-prototype is made. The next step of the work is to conduct research activities both in the field of dynamics of platform parallel kinematics manipulators with six degrees of freedom and in the

  14. Efficient Parallel Algorithms for Landscape Evolution Modelling

    Science.gov (United States)

    Moresi, L. N.; Mather, B.; Beucher, R.

    2017-12-01

    Landscape erosion and the deposition of sediments by river systems are strongly controlled bytopography, rainfall patterns, and the susceptibility of the basement to the action ofrunning water. It is well understood that each of these processes depends on the other, for example:topography results from active tectonic processes; deformation, metamorphosis andexhumation alter the competence of the basement; rainfall patterns depend on topography;uplift and subsidence in response to tectonic stress can be amplified by erosionand sediment deposition. We typically gain understanding of such coupled systems through forward models which capture theessential interactions of the various components and attempt parameterise those parts of the individual systemthat are unresolvable at the scale of the interaction. Here we address the problem of predicting erosion and deposition rates at a continental scalewith a resolution of tens to hundreds of metres in a dynamic, Lagrangian framework. This isa typical requirement for a code to interface with a mantle / lithosphere dynamics model anddemands an efficient, unstructured, parallel implementation. We address this through a very general algorithm that treats all parts of the landscape evolution equationsin sparse-matrix form including those for stream-flow accumulation, dam-filling and catchment determination. This givesus considerable flexibility in developing unstructured, parallel code, and in creating a modular packagethat can be configured by users to work at different temporal and spatial scales, but is also has potential advantagesin treating the non-linear parts of the problem in a general manner.

  15. Data mining with SPSS modeler theory, exercises and solutions

    CERN Document Server

    Wendler, Tilo

    2016-01-01

    Introducing the IBM SPSS Modeler, this book guides readers through data mining processes and presents relevant statistical methods. There is a special focus on step-by-step tutorials and well-documented examples that help demystify complex mathematical algorithms and computer programs. The variety of exercises and solutions as well as an accompanying website with data sets and SPSS Modeler streams are particularly valuable. While intended for students, the simplicity of the Modeler makes the book useful for anyone wishing to learn about basic and more advanced data mining, and put this knowledge into practice.

  16. Focuss algorithm application in kinetic compartment modeling for PET tracer

    International Nuclear Information System (INIS)

    Huang Xinrui; Bao Shanglian

    2004-01-01

    Molecular imaging is in the process of becoming. Its application mostly depends on the molecular discovery process of imaging probes and drugs, from the mouse to the patient, from research to clinical practice. Positron emission tomography (PET) can non-invasively monitor . pharmacokinetic and functional processes of drugs in intact organisms at tracer concentrations by kinetic modeling. It has been known that for all biological systems, linear or nonlinear, if the system is injected by a tracer in a steady state, the distribution of the tracer follows the kinetics of a linear compartmental system, which has sums of exponential solutions. Based on the general compartmental description of the tracer's fate in vivo, we presented a novel kinetic modeling approach for the quantification of in vivo tracer studies with dynamic positron emission tomography (PET), which can determine a parsimonious model consisting with the measured data. This kinetic modeling technique allows for estimation of parametric images from a voxel based analysis and requires no a priori decision about the tracer's fate in vivo, instead determining the most appropriate model from the information contained within the kinetic data. Choosing a set of exponential functions, convolved with the plasma input function, as basis functions, the time activity curve of a region or a pixel can be written as a linear combination of the basis functions with corresponding coefficients. The number of non-zero coefficients returned corresponds to the model order which is related to the number of tissue compartments. The system macro parameters are simply determined using the focal underdetermined system solver (FOCUSS) algorithm. The FOCUSS algorithm is a nonparametric algorithm for finding localized energy solutions from limited data and is a recursive linear estimation procedure. FOCUSS algorithm usually converges very fast, so demands a few iterations. The effectiveness is verified by simulation and clinical

  17. Portfolio optimization by using linear programing models based on genetic algorithm

    Science.gov (United States)

    Sukono; Hidayat, Y.; Lesmana, E.; Putra, A. S.; Napitupulu, H.; Supian, S.

    2018-01-01

    In this paper, we discussed the investment portfolio optimization using linear programming model based on genetic algorithms. It is assumed that the portfolio risk is measured by absolute standard deviation, and each investor has a risk tolerance on the investment portfolio. To complete the investment portfolio optimization problem, the issue is arranged into a linear programming model. Furthermore, determination of the optimum solution for linear programming is done by using a genetic algorithm. As a numerical illustration, we analyze some of the stocks traded on the capital market in Indonesia. Based on the analysis, it is shown that the portfolio optimization performed by genetic algorithm approach produces more optimal efficient portfolio, compared to the portfolio optimization performed by a linear programming algorithm approach. Therefore, genetic algorithms can be considered as an alternative on determining the investment portfolio optimization, particularly using linear programming models.

  18. Model order reduction using eigen algorithm

    African Journals Online (AJOL)

    DR OKE

    to use either for design or analysis. Hence, it is ... directly from the Eigen algorithm while the zeros are determined through factor division algorithm to obtain the reduced order system. ..... V. Singh, Chandra and H. Kar, “Improved Routh Pade approximationss: A computer aided approach”, IEEE Transaction on. Automat ...

  19. Routing Flow-Shop with Buffers and Ready Times – Comparison of Selected Solution Algorithms

    Directory of Open Access Journals (Sweden)

    Józefczyk Jerzy

    2014-12-01

    Full Text Available This article extends the former results concerning the routing flow-shop problem to minimize the makespan on the case with buffers, non-zero ready times and different speeds of machines. The corresponding combinatorial optimization problem is formulated. The exact as well as four heuristic solution algorithms are presented. The branch and bound approach is applied for the former one. The heuristic algorithms employ known constructive idea proposed for the former version of the problem as well as the Tabu Search metaheuristics. Moreover, the improvement procedure is proposed to enhance the quality of both heuristic algorithms. The conducted simulation experiments allow evaluating all algorithms. Firstly, the heuristic algorithms are compared with the exact one for small instances of the problem in terms of the criterion and execution times. Then, for larger instances, the heuristic algorithms are mutually compared. The case study regarding the maintenance of software products, given in the final part of the paper, illustrates the possibility to apply the results for real-world manufacturing systems.

  20. Algorithm Development for the Two-Fluid Plasma Model

    National Research Council Canada - National Science Library

    Shumlak, Uri

    2002-01-01

    A preliminary algorithm based on the two-fluid plasma model is developed to investigate the possibility of simulating plasmas with a more physically accurate model than the MHD (magnetohydrodynamic) model...

  1. Pseudoclassical fermionic model and classical solutions

    International Nuclear Information System (INIS)

    Smailagic, A.

    1981-08-01

    We study classical limit of fermionic fields seen as Grassmann variables and deduce the proper quantization prescription using Dirac's method for constrained systems and investigate quantum meaning of classical solutions for the Thirring model. (author)

  2. Convergence of an Iterative Algorithm for Common Solutions for Zeros of Maximal Accretive Operator with Applications

    Directory of Open Access Journals (Sweden)

    Uamporn Witthayarat

    2012-01-01

    Full Text Available The aim of this paper is to introduce an iterative algorithm for finding a common solution of the sets (A+M2−1(0 and (B+M1−1(0, where M is a maximal accretive operator in a Banach space and, by using the proposed algorithm, to establish some strong convergence theorems for common solutions of the two sets above in a uniformly convex and 2-uniformly smooth Banach space. The results obtained in this paper extend and improve the corresponding results of Qin et al. 2011 from Hilbert spaces to Banach spaces and Petrot et al. 2011. Moreover, we also apply our results to some applications for solving convex feasibility problems.

  3. A finite element solution algorithm for the Navier-Stokes equations

    Science.gov (United States)

    Baker, A. J.

    1974-01-01

    A finite element solution algorithm is established for the two-dimensional Navier-Stokes equations governing the steady-state kinematics and thermodynamics of a variable viscosity, compressible multiple-species fluid. For an incompressible fluid, the motion may be transient as well. The primitive dependent variables are replaced by a vorticity-streamfunction description valid in domains spanned by rectangular, cylindrical and spherical coordinate systems. Use of derived variables provides a uniformly elliptic partial differential equation description for the Navier-Stokes system, and for which the finite element algorithm is established. Explicit non-linearity is accepted by the theory, since no psuedo-variational principles are employed, and there is no requirement for either computational mesh or solution domain closure regularity. Boundary condition constraints on the normal flux and tangential distribution of all computational variables, as well as velocity, are routinely piecewise enforceable on domain closure segments arbitrarily oriented with respect to a global reference frame.

  4. Investigation of ALEGRA shock hydrocode algorithms using an exact free surface jet flow solution.

    Energy Technology Data Exchange (ETDEWEB)

    Hanks, Bradley Wright.; Robinson, Allen C

    2014-01-01

    Computational testing of the arbitrary Lagrangian-Eulerian shock physics code, ALEGRA, is presented using an exact solution that is very similar to a shaped charge jet flow. The solution is a steady, isentropic, subsonic free surface flow with significant compression and release and is provided as a steady state initial condition. There should be no shocks and no entropy production throughout the problem. The purpose of this test problem is to present a detailed and challenging computation in order to provide evidence for algorithmic strengths and weaknesses in ALEGRA which should be examined further. The results of this work are intended to be used to guide future algorithmic improvements in the spirit of test-driven development processes.

  5. A Formal Verification Model for Performance Analysis of Reinforcement Learning Algorithms Applied t o Dynamic Networks

    OpenAIRE

    Shrirang Ambaji KULKARNI; Raghavendra G . RAO

    2017-01-01

    Routing data packets in a dynamic network is a difficult and important problem in computer networks. As the network is dynamic, it is subject to frequent topology changes and is subject to variable link costs due to congestion and bandwidth. Existing shortest path algorithms fail to converge to better solutions under dynamic network conditions. Reinforcement learning algorithms posses better adaptation techniques in dynamic environments. In this paper we apply model based Q-Routing technique ...

  6. Dynamics analysis of electrodynamic satellite tethers. Equations of motion and numerical solution algorithms for the tether

    Science.gov (United States)

    Nacozy, P. E.

    1984-01-01

    The equations of motion are developed for a perfectly flexible, inelastic tether with a satellite at its extremity. The tether is attached to a space vehicle in orbit. The tether is allowed to possess electrical conductivity. A numerical solution algorithm to provide the motion of the tether and satellite system is presented. The resulting differential equations can be solved by various existing standard numerical integration computer programs. The resulting differential equations allow the introduction of approximations that can lead to analytical, approximate general solutions. The differential equations allow more dynamical insight of the motion.

  7. Solution Algorithm for a New Bi-Level Discrete Network Design Problem

    Directory of Open Access Journals (Sweden)

    Qun Chen

    2013-12-01

    Full Text Available A new discrete network design problem (DNDP was pro-posed in this paper, where the variables can be a series of integers rather than just 0-1. The new DNDP can determine both capacity improvement grades of reconstruction roads and locations and capacity grades of newly added roads, and thus complies with the practical projects where road capacity can only be some discrete levels corresponding to the number of lanes of roads. This paper designed a solution algorithm combining branch-and-bound with Hooke-Jeeves algorithm, where feasible integer solutions are recorded in searching the process of Hooke-Jeeves algorithm, lend -ing itself to determine the upper bound of the upper-level problem. The thresholds for branch cutting and ending were set for earlier convergence. Numerical examples are given to demonstrate the efficiency of the proposed algorithm.

  8. An algorithm for computing the hull of the solution set of interval linear equations

    Czech Academy of Sciences Publication Activity Database

    Rohn, Jiří

    2011-01-01

    Roč. 435, č. 2 (2011), s. 193-201 ISSN 0024-3795 R&D Projects: GA ČR GA201/09/1957; GA ČR GC201/08/J020 Institutional research plan: CEZ:AV0Z10300504 Keywords : interval linear equations * solution set * interval hull * algorithm * absolute value inequality Subject RIV: BA - General Mathematics Impact factor: 0.974, year: 2011

  9. Efficient Implementation Algorithms for Homogenized Energy Models

    National Research Council Canada - National Science Library

    Braun, Thomas R; Smith, Ralph C

    2005-01-01

    ... for real-time control implementation. In this paper, we develop algorithms employing lookup tables which permit the high speed implementation of formulations which incorporate relaxation mechanisms and electromechanical coupling...

  10. KnowBaSICS-M: an ontology-based system for semantic management of medical problems and computerised algorithmic solutions.

    Science.gov (United States)

    Bratsas, Charalampos; Koutkias, Vassilis; Kaimakamis, Evangelos; Bamidis, Panagiotis D; Pangalos, George I; Maglaveras, Nicos

    2007-10-01

    In this paper, an ontology-based system (KnowBaSICS-M) is presented for the semantic management of Medical Computational Problems (MCPs), i.e., medical problems and computerised algorithmic solutions. The system provides an open environment, which: (1) allows clinicians and researchers to retrieve potential algorithmic solutions pertinent to a medical problem and (2) enables incorporation of new MCPs into its underlying Knowledge Base (KB). KnowBaSICS-M is a modular system for MCP acquisition and discovery that relies on an innovative ontology-based model incorporating concepts from the Unified Medical Language System (UMLS). Information retrieval (IR) is based on an ontology-based Vector Space Model (VSM) that estimates the similarity among user-defined MCP search criteria and registered MCP solutions in the KB. The results of a preliminary evaluation and specific examples of use are presented to illustrate the benefits of the system. KnowBaSICS-M constitutes an approach towards the construction of an integrated and manageable MCP repository for the biomedical research community.

  11. The quantum Rabi model: solution and dynamics

    International Nuclear Information System (INIS)

    Xie, Qiongtao; Zhong, Honghua; Lee, Chaohong; Batchelor, Murray T

    2017-01-01

    This article presents a review of recent developments on various aspects of the quantum Rabi model. Particular emphasis is given on the exact analytic solution obtained in terms of confluent Heun functions. The analytic solutions for various generalisations of the quantum Rabi model are also discussed. Results are also reviewed on the level statistics and the dynamics of the quantum Rabi model. The article concludes with an introductory overview of several experimental realisations of the quantum Rabi model. An outlook towards future developments is also given. (topical review)

  12. Loop algorithms for quantum simulations of fermion models on lattices

    International Nuclear Information System (INIS)

    Kawashima, N.; Gubernatis, J.E.; Evertz, H.G.

    1994-01-01

    Two cluster algorithms, based on constructing and flipping loops, are presented for world-line quantum Monte Carlo simulations of fermions and are tested on the one-dimensional repulsive Hubbard model. We call these algorithms the loop-flip and loop-exchange algorithms. For these two algorithms and the standard world-line algorithm, we calculated the autocorrelation times for various physical quantities and found that the ordinary world-line algorithm, which uses only local moves, suffers from very long correlation times that makes not only the estimate of the error difficult but also the estimate of the average values themselves difficult. These difficulties are especially severe in the low-temperature, large-U regime. In contrast, we find that new algorithms, when used alone or in combinations with themselves and the standard algorithm, can have significantly smaller autocorrelation times, in some cases being smaller by three orders of magnitude. The new algorithms, which use nonlocal moves, are discussed from the point of view of a general prescription for developing cluster algorithms. The loop-flip algorithm is also shown to be ergodic and to belong to the grand canonical ensemble. Extensions to other models and higher dimensions are briefly discussed

  13. Optimizing the Forward Algorithm for Hidden Markov Model on IBM Roadrunner clusters

    Directory of Open Access Journals (Sweden)

    SOIMAN, S.-I.

    2015-05-01

    Full Text Available In this paper we present a parallel solution of the Forward Algorithm for Hidden Markov Models. The Forward algorithm compute a probability of a hidden state from Markov model at a certain time, this process being recursively. The whole process requires large computational resources for those models with a large number of states and long observation sequences. Our solution in order to reduce the computational time is a multilevel parallelization of Forward algorithm. Two types of cores were used in our implementation, for each level of parallelization, cores that are graved on the same chip of PowerXCell8i processor. This hybrid architecture of processors permitted us to obtain a speedup factor over 40 relative to the sequential algorithm for a model with 24 states and 25 millions of observable symbols. Experimental results showed that the parallel Forward algorithm can evaluate the probability of an observation sequence on a hidden Markov model 40 times faster than the classic one does. Based on the performance obtained, we demonstrate the applicability of this parallel implementation of Forward algorithm in complex problems such as large vocabulary speech recognition.

  14. The Crystal-T algorithm: a new approach to calculate the SLE of lipidic mixtures presenting solid solutions.

    Science.gov (United States)

    Maximo, Guilherme J; Costa, Mariana C; Meirelles, Antonio J A

    2014-08-21

    Lipidic mixtures present a particular phase change profile highly affected by their unique crystalline structure. However, classical solid-liquid equilibrium (SLE) thermodynamic modeling approaches, which assume the solid phase to be a pure component, sometimes fail in the correct description of the phase behavior. In addition, their inability increases with the complexity of the system. To overcome some of these problems, this study describes a new procedure to depict the SLE of fatty binary mixtures presenting solid solutions, namely the "Crystal-T algorithm". Considering the non-ideality of both liquid and solid phases, this algorithm is aimed at the determination of the temperature in which the first and last crystal of the mixture melts. The evaluation is focused on experimental data measured and reported in this work for systems composed of triacylglycerols and fatty alcohols. The liquidus and solidus lines of the SLE phase diagrams were described by using excess Gibbs energy based equations, and the group contribution UNIFAC model for the calculation of the activity coefficients of both liquid and solid phases. Very low deviations of theoretical and experimental data evidenced the strength of the algorithm, contributing to the enlargement of the scope of the SLE modeling.

  15. GRAVITATIONAL LENS MODELING WITH GENETIC ALGORITHMS AND PARTICLE SWARM OPTIMIZERS

    International Nuclear Information System (INIS)

    Rogers, Adam; Fiege, Jason D.

    2011-01-01

    Strong gravitational lensing of an extended object is described by a mapping from source to image coordinates that is nonlinear and cannot generally be inverted analytically. Determining the structure of the source intensity distribution also requires a description of the blurring effect due to a point-spread function. This initial study uses an iterative gravitational lens modeling scheme based on the semilinear method to determine the linear parameters (source intensity profile) of a strongly lensed system. Our 'matrix-free' approach avoids construction of the lens and blurring operators while retaining the least-squares formulation of the problem. The parameters of an analytical lens model are found through nonlinear optimization by an advanced genetic algorithm (GA) and particle swarm optimizer (PSO). These global optimization routines are designed to explore the parameter space thoroughly, mapping model degeneracies in detail. We develop a novel method that determines the L-curve for each solution automatically, which represents the trade-off between the image χ 2 and regularization effects, and allows an estimate of the optimally regularized solution for each lens parameter set. In the final step of the optimization procedure, the lens model with the lowest χ 2 is used while the global optimizer solves for the source intensity distribution directly. This allows us to accurately determine the number of degrees of freedom in the problem to facilitate comparison between lens models and enforce positivity on the source profile. In practice, we find that the GA conducts a more thorough search of the parameter space than the PSO.

  16. The Evolutionary Algorithm to Find Robust Pareto-Optimal Solutions over Time

    Directory of Open Access Journals (Sweden)

    Meirong Chen

    2015-01-01

    Full Text Available In dynamic multiobjective optimization problems, the environmental parameters change over time, which makes the true pareto fronts shifted. So far, most works of research on dynamic multiobjective optimization methods have concentrated on detecting the changed environment and triggering the population based optimization methods so as to track the moving pareto fronts over time. Yet, in many real-world applications, it is not necessary to find the optimal nondominant solutions in each dynamic environment. To solve this weakness, a novel method called robust pareto-optimal solution over time is proposed. It is in fact to replace the optimal pareto front at each time-varying moment with the series of robust pareto-optimal solutions. This means that each robust solution can fit for more than one time-varying moment. Two metrics, including the average survival time and average robust generational distance, are present to measure the robustness of the robust pareto solution set. Another contribution is to construct the algorithm framework searching for robust pareto-optimal solutions over time based on the survival time. Experimental results indicate that this definition is a more practical and time-saving method of addressing dynamic multiobjective optimization problems changing over time.

  17. An Evolutionary Algorithm for Multiobjective Fuzzy Portfolio Selection Models with Transaction Cost and Liquidity

    Directory of Open Access Journals (Sweden)

    Wei Yue

    2015-01-01

    Full Text Available The major issues for mean-variance-skewness models are the errors in estimations that cause corner solutions and low diversity in the portfolio. In this paper, a multiobjective fuzzy portfolio selection model with transaction cost and liquidity is proposed to maintain the diversity of portfolio. In addition, we have designed a multiobjective evolutionary algorithm based on decomposition of the objective space to maintain the diversity of obtained solutions. The algorithm is used to obtain a set of Pareto-optimal portfolios with good diversity and convergence. To demonstrate the effectiveness of the proposed model and algorithm, the performance of the proposed algorithm is compared with the classic MOEA/D and NSGA-II through some numerical examples based on the data of the Shanghai Stock Exchange Market. Simulation results show that our proposed algorithm is able to obtain better diversity and more evenly distributed Pareto front than the other two algorithms and the proposed model can maintain quite well the diversity of portfolio. The purpose of this paper is to deal with portfolio problems in the weighted possibilistic mean-variance-skewness (MVS and possibilistic mean-variance-skewness-entropy (MVS-E frameworks with transaction cost and liquidity and to provide different Pareto-optimal investment strategies as diversified as possible for investors at a time, rather than one strategy for investors at a time.

  18. Variable selection in Logistic regression model with genetic algorithm.

    Science.gov (United States)

    Zhang, Zhongheng; Trevino, Victor; Hoseini, Sayed Shahabuddin; Belciug, Smaranda; Boopathi, Arumugam Manivanna; Zhang, Ping; Gorunescu, Florin; Subha, Velappan; Dai, Songshi

    2018-02-01

    Variable or feature selection is one of the most important steps in model specification. Especially in the case of medical-decision making, the direct use of a medical database, without a previous analysis and preprocessing step, is often counterproductive. In this way, the variable selection represents the method of choosing the most relevant attributes from the database in order to build a robust learning models and, thus, to improve the performance of the models used in the decision process. In biomedical research, the purpose of variable selection is to select clinically important and statistically significant variables, while excluding unrelated or noise variables. A variety of methods exist for variable selection, but none of them is without limitations. For example, the stepwise approach, which is highly used, adds the best variable in each cycle generally producing an acceptable set of variables. Nevertheless, it is limited by the fact that it commonly trapped in local optima. The best subset approach can systematically search the entire covariate pattern space, but the solution pool can be extremely large with tens to hundreds of variables, which is the case in nowadays clinical data. Genetic algorithms (GA) are heuristic optimization approaches and can be used for variable selection in multivariable regression models. This tutorial paper aims to provide a step-by-step approach to the use of GA in variable selection. The R code provided in the text can be extended and adapted to other data analysis needs.

  19. Aqueous Solution Vessel Thermal Model Development II

    Energy Technology Data Exchange (ETDEWEB)

    Buechler, Cynthia Eileen [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)

    2015-10-28

    The work presented in this report is a continuation of the work described in the May 2015 report, “Aqueous Solution Vessel Thermal Model Development”. This computational fluid dynamics (CFD) model aims to predict the temperature and bubble volume fraction in an aqueous solution of uranium. These values affect the reactivity of the fissile solution, so it is important to be able to calculate them and determine their effects on the reaction. Part A of this report describes some of the parameter comparisons performed on the CFD model using Fluent. Part B describes the coupling of the Fluent model with a Monte-Carlo N-Particle (MCNP) neutron transport model. The fuel tank geometry is the same as it was in the May 2015 report, annular with a thickness-to-height ratio of 0.16. An accelerator-driven neutron source provides the excitation for the reaction, and internal and external water cooling channels remove the heat. The model used in this work incorporates the Eulerian multiphase model with lift, wall lubrication, turbulent dispersion and turbulence interaction. The buoyancy-driven flow is modeled using the Boussinesq approximation, and the flow turbulence is determined using the k-ω Shear-Stress-Transport (SST) model. The dispersed turbulence multiphase model is employed to capture the multiphase turbulence effects.

  20. PARTRACK - A particle tracking algorithm for transport and dispersion of solutes in a sparsely fractured rock

    International Nuclear Information System (INIS)

    Svensson, Urban

    2001-04-01

    A particle tracking algorithm, PARTRACK, that simulates transport and dispersion in a sparsely fractured rock is described. The main novel feature of the algorithm is the introduction of multiple particle states. It is demonstrated that the introduction of this feature allows for the simultaneous simulation of Taylor dispersion, sorption and matrix diffusion. A number of test cases are used to verify and demonstrate the features of PARTRACK. It is shown that PARTRACK can simulate the following processes, believed to be important for the problem addressed: the split up of a tracer cloud at a fracture intersection, channeling in a fracture plane, Taylor dispersion and matrix diffusion and sorption. From the results of the test cases, it is concluded that PARTRACK is an adequate framework for simulation of transport and dispersion of a solute in a sparsely fractured rock

  1. Application of stochastic weighted algorithms to a multidimensional silica particle model

    Energy Technology Data Exchange (ETDEWEB)

    Menz, William J. [Department of Chemical Engineering and Biotechnology, University of Cambridge, New Museums Site, Pembroke Street, Cambridge CB2 3RA (United Kingdom); Patterson, Robert I.A.; Wagner, Wolfgang [Weierstrass Institute for Applied Analysis and Stochastics, Mohrenstrasse 39, Berlin 10117 (Germany); Kraft, Markus, E-mail: mk306@cam.ac.uk [Department of Chemical Engineering and Biotechnology, University of Cambridge, New Museums Site, Pembroke Street, Cambridge CB2 3RA (United Kingdom)

    2013-09-01

    Highlights: •Stochastic weighted algorithms (SWAs) are developed for a detailed silica model. •An implementation of SWAs with the transition kernel is presented. •The SWAs’ solutions converge to the direct simulation algorithm’s (DSA) solution. •The efficiency of SWAs is evaluated for this multidimensional particle model. •It is shown that SWAs can be used for coagulation problems in industrial systems. -- Abstract: This paper presents a detailed study of the numerical behaviour of stochastic weighted algorithms (SWAs) using the transition regime coagulation kernel and a multidimensional silica particle model. The implementation in the SWAs of the transition regime coagulation kernel and associated majorant rates is described. The silica particle model of Shekar et al. [S. Shekar, A.J. Smith, W.J. Menz, M. Sander, M. Kraft, A multidimensional population balance model to describe the aerosol synthesis of silica nanoparticles, Journal of Aerosol Science 44 (2012) 83–98] was used in conjunction with this coagulation kernel to study the convergence properties of SWAs with a multidimensional particle model. High precision solutions were calculated with two SWAs and also with the established direct simulation algorithm. These solutions, which were generated using large number of computational particles, showed close agreement. It was thus demonstrated that SWAs can be successfully used with complex coagulation kernels and high dimensional particle models to simulate real-world systems.

  2. Methodology and basic algorithms of the Livermore Economic Modeling System

    Energy Technology Data Exchange (ETDEWEB)

    Bell, R.B.

    1981-03-17

    The methodology and the basic pricing algorithms used in the Livermore Economic Modeling System (EMS) are described. The report explains the derivations of the EMS equations in detail; however, it could also serve as a general introduction to the modeling system. A brief but comprehensive explanation of what EMS is and does, and how it does it is presented. The second part examines the basic pricing algorithms currently implemented in EMS. Each algorithm's function is analyzed and a detailed derivation of the actual mathematical expressions used to implement the algorithm is presented. EMS is an evolving modeling system; improvements in existing algorithms are constantly under development and new submodels are being introduced. A snapshot of the standard version of EMS is provided and areas currently under study and development are considered briefly.

  3. Mathematical Model of Suspension Filtration and Its Analytical Solution

    Directory of Open Access Journals (Sweden)

    Normahmad Ravshanov

    2013-01-01

    Full Text Available The work develops advanced mathematical model and computing algorithm to analyze, predict and identify the basic parameters of filter units and their variation ranges. Numerical analytic solution of liquid ionized mixtures filtration was got on their basis. Computing experiments results are presented in graphics form. Calculation results analysis enables to determine the optimum performance of filter units, used for liquid ionized mixtures filtration, food preparation, drug production and water purification. Selection of the most suitable parameters contributes to the improvement of economic and technological efficiency of production and filter units working efficiency.

  4. Comparison of parameter estimation algorithms in hydrological modelling

    DEFF Research Database (Denmark)

    Blasone, Roberta-Serena; Madsen, Henrik; Rosbjerg, Dan

    2006-01-01

    for these types of models, although at a more expensive computational cost. The main purpose of this study is to investigate the performance of a global and a local parameter optimization algorithm, respectively, the Shuffled Complex Evolution (SCE) algorithm and the gradient-based Gauss...

  5. Evaluation of models generated via hybrid evolutionary algorithms ...

    African Journals Online (AJOL)

    2016-04-02

    Apr 2, 2016 ... Evaluation of models generated via hybrid evolutionary algorithms for the prediction of Microcystis ... evolutionary algorithms (HEA) proved to be highly applica- ble to the hypertrophic reservoirs of South Africa. .... discovered and optimised using a large-scale parallel computational device and relevant soft-.

  6. Algorithms

    Indian Academy of Sciences (India)

    have been found in Vedic Mathematics which are dated much before Euclid's algorithm. A programming language Is used to describe an algorithm for execution on a computer. An algorithm expressed using a programming language Is called a program. From activities 1-3, we can observe that: • Each activity is a command.

  7. Insertion algorithms for network model database management systems

    Science.gov (United States)

    Mamadolimov, Abdurashid; Khikmat, Saburov

    2017-12-01

    The network model is a database model conceived as a flexible way of representing objects and their relationships. Its distinguishing feature is that the schema, viewed as a graph in which object types are nodes and relationship types are arcs, forms partial order. When a database is large and a query comparison is expensive then the efficiency requirement of managing algorithms is minimizing the number of query comparisons. We consider updating operation for network model database management systems. We develop a new sequantial algorithm for updating operation. Also we suggest a distributed version of the algorithm.

  8. Ionospheric Scintillation Effects on GPS Measurements and Algorithms to Improve Positioning Solution Accuracy

    Science.gov (United States)

    Myer, Gregory Thomas

    The ionosphere is an important cause of disturbances on GNSS signals, especially in high latitudes and equatorial areas. Previous studies indicate that while ionospheric scintillation may cause abrupt, random fluctuations in carrier phase measurements, its impact on pseu- dorange is less serious. Since modern GNSS receivers, especially those for high precision applications, use carrier phase-smoothed pseudoranges to improve accuracy of position solutions, there exists the need to have a better understanding of the scintillation effects on carrier phase measurements and developing means to mitigate scintillation induced errors in navigation solutions. In this thesis, scintillation impacts are demonstrated on carrier phase and pseudorange measurements using real scintillation data collected at high latitudes and equatorial areas, and the effect on positioning is investigated and mitigated. To obtain a more insightful and quantitative understanding of the impact, the data was used to generate position solutions using standard navigation processing algorithms. The results clearly indicate that sudden carrier phase discontinuities during strong scintillation lead to the degradation of carrier-smoothed pseudorange accuracy and consequently, results in large position errors. During strong scintillation with no carrier phase discontinuities, comparatively smaller position er- rors are found due to phase fluctuations that cause small changes in the range measurements. Based on this analysis, we give examples of several approaches to mitigate these problems, and use these approaches to present adaptive positioning techniques to mitigate scintillation induced position errors. One algorithm simply replaces the carrier-smoothed pseudorange with the unsmoothed pseudorange for satellites that are affected by outages on the carrier phase measurements, or if strong scintillation is detected. Another adaptive algorithm uses the GDOP to determine if a scintillating satellite can be

  9. Analyzing Traffic Problem Model With Graph Theory Algorithms

    OpenAIRE

    Tan, Yong

    2014-01-01

    This paper will contribute to a practical problem, Urban Traffic. We will investigate those features, try to simplify the complexity and formulize this dynamic system. These contents mainly contain how to analyze a decision problem with combinatorial method and graph theory algorithms; how to optimize our strategy to gain a feasible solution through employing other principles of Computer Science.

  10. An Automatic Registration Algorithm for 3D Maxillofacial Model

    Science.gov (United States)

    Qiu, Luwen; Zhou, Zhongwei; Guo, Jixiang; Lv, Jiancheng

    2016-09-01

    3D image registration aims at aligning two 3D data sets in a common coordinate system, which has been widely used in computer vision, pattern recognition and computer assisted surgery. One challenging problem in 3D registration is that point-wise correspondences between two point sets are often unknown apriori. In this work, we develop an automatic algorithm for 3D maxillofacial models registration including facial surface model and skull model. Our proposed registration algorithm can achieve a good alignment result between partial and whole maxillofacial model in spite of ambiguous matching, which has a potential application in the oral and maxillofacial reparative and reconstructive surgery. The proposed algorithm includes three steps: (1) 3D-SIFT features extraction and FPFH descriptors construction; (2) feature matching using SAC-IA; (3) coarse rigid alignment and refinement by ICP. Experiments on facial surfaces and mandible skull models demonstrate the efficiency and robustness of our algorithm.

  11. Algorithmic detectability threshold of the stochastic block model

    Science.gov (United States)

    Kawamoto, Tatsuro

    2018-03-01

    The assumption that the values of model parameters are known or correctly learned, i.e., the Nishimori condition, is one of the requirements for the detectability analysis of the stochastic block model in statistical inference. In practice, however, there is no example demonstrating that we can know the model parameters beforehand, and there is no guarantee that the model parameters can be learned accurately. In this study, we consider the expectation-maximization (EM) algorithm with belief propagation (BP) and derive its algorithmic detectability threshold. Our analysis is not restricted to the community structure but includes general modular structures. Because the algorithm cannot always learn the planted model parameters correctly, the algorithmic detectability threshold is qualitatively different from the one with the Nishimori condition.

  12. Genetic Algorithms for a Parameter Estimation of a Fermentation Process Model: A Comparison

    Directory of Open Access Journals (Sweden)

    Olympia Roeva

    2005-12-01

    Full Text Available In this paper the problem of a parameter estimation using genetic algorithms is examined. A case study considering the estimation of 6 parameters of a nonlinear dynamic model of E. coli fermentation is presented as a test problem. The parameter estimation problem is stated as a nonlinear programming problem subject to nonlinear differential-algebraic constraints. This problem is known to be frequently ill-conditioned and multimodal. Thus, traditional (gradient-based local optimization methods fail to arrive satisfied solutions. To overcome their limitations, the use of different genetic algorithms as stochastic global optimization methods is explored. These algorithms are proved to be very suitable for the optimization of highly non-linear problems with many variables. Genetic algorithms can guarantee global optimality and robustness. These facts make them advantageous in use for parameter identification of fermentation models. A comparison between simple, modified and multi-population genetic algorithms is presented. The best result is obtained using the modified genetic algorithm. The considered algorithms converged very closely to the cost value but the modified algorithm is in times faster than other two.

  13. Fast integral equation algorithms for the solution of electromagnetic wave propagation over general terrains

    Directory of Open Access Journals (Sweden)

    Ibrahim K. Abu Seif

    2015-01-01

    Full Text Available In this paper a fast numerical algorithm to solve an integral equation model for wave propagation along a perfectly conducting two-dimensional terrain is suggested. It is applied to different actual terrain profiles and the results indicate very good agreement with published work. In addition, the proposed algorithm has achieved considerable saving in processing time. The formulation is extended to solve the propagation over lossy dielectric surfaces. A combined field integral equation (CFIE for wave propagation over dielectric terrain is solved efficiently by utilizing the method of moments with complex basis functions. The numerical results for different cases of dielectric surfaces are compared with the results of perfectly conducting surface evaluated by the IE conventional algorithm.

  14. An analysis dictionary learning algorithm under a noisy data model with orthogonality constraint.

    Science.gov (United States)

    Zhang, Ye; Yu, Tenglong; Wang, Wenwu

    2014-01-01

    Two common problems are often encountered in analysis dictionary learning (ADL) algorithms. The first one is that the original clean signals for learning the dictionary are assumed to be known, which otherwise need to be estimated from noisy measurements. This, however, renders a computationally slow optimization process and potentially unreliable estimation (if the noise level is high), as represented by the Analysis K-SVD (AK-SVD) algorithm. The other problem is the trivial solution to the dictionary, for example, the null dictionary matrix that may be given by a dictionary learning algorithm, as discussed in the learning overcomplete sparsifying transform (LOST) algorithm. Here we propose a novel optimization model and an iterative algorithm to learn the analysis dictionary, where we directly employ the observed data to compute the approximate analysis sparse representation of the original signals (leading to a fast optimization procedure) and enforce an orthogonality constraint on the optimization criterion to avoid the trivial solutions. Experiments demonstrate the competitive performance of the proposed algorithm as compared with three baselines, namely, the AK-SVD, LOST, and NAAOLA algorithms.

  15. An Analysis Dictionary Learning Algorithm under a Noisy Data Model with Orthogonality Constraint

    Directory of Open Access Journals (Sweden)

    Ye Zhang

    2014-01-01

    Full Text Available Two common problems are often encountered in analysis dictionary learning (ADL algorithms. The first one is that the original clean signals for learning the dictionary are assumed to be known, which otherwise need to be estimated from noisy measurements. This, however, renders a computationally slow optimization process and potentially unreliable estimation (if the noise level is high, as represented by the Analysis K-SVD (AK-SVD algorithm. The other problem is the trivial solution to the dictionary, for example, the null dictionary matrix that may be given by a dictionary learning algorithm, as discussed in the learning overcomplete sparsifying transform (LOST algorithm. Here we propose a novel optimization model and an iterative algorithm to learn the analysis dictionary, where we directly employ the observed data to compute the approximate analysis sparse representation of the original signals (leading to a fast optimization procedure and enforce an orthogonality constraint on the optimization criterion to avoid the trivial solutions. Experiments demonstrate the competitive performance of the proposed algorithm as compared with three baselines, namely, the AK-SVD, LOST, and NAAOLA algorithms.

  16. Implementation of a Multichannel Serial Data Streaming Algorithm using the Xilinx Serial RapidIO Solution

    Science.gov (United States)

    Doxley, Charles A.

    2016-01-01

    In the current world of applications that use reconfigurable technology implemented on field programmable gate arrays (FPGAs), there is a need for flexible architectures that can grow as the systems evolve. A project has limited resources and a fixed set of requirements that development efforts are tasked to meet. Designers must develop robust solutions that practically meet the current customer demands and also have the ability to grow for future performance. This paper describes the development of a high speed serial data streaming algorithm that allows for transmission of multiple data channels over a single serial link. The technique has the ability to change to meet new applications developed for future design considerations. This approach uses the Xilinx Serial RapidIO LOGICORE Solution to implement a flexible infrastructure to meet the current project requirements with the ability to adapt future system designs.

  17. Introduction of Parallel GPGPU Acceleration Algorithms for the Solution of Radiative Transfer

    Science.gov (United States)

    Godoy, William F.; Liu, Xu

    2011-01-01

    General-purpose computing on graphics processing units (GPGPU) is a recent technique that allows the parallel graphics processing unit (GPU) to accelerate calculations performed sequentially by the central processing unit (CPU). To introduce GPGPU to radiative transfer, the Gauss-Seidel solution of the well-known expressions for 1-D and 3-D homogeneous, isotropic media is selected as a test case. Different algorithms are introduced to balance memory and GPU-CPU communication, critical aspects of GPGPU. Results show that speed-ups of one to two orders of magnitude are obtained when compared to sequential solutions. The underlying value of GPGPU is its potential extension in radiative solvers (e.g., Monte Carlo, discrete ordinates) at a minimal learning curve.

  18. Accelerated Genetic Algorithm Solutions Of Some Parametric Families Of Stochastic Differential Equations

    Directory of Open Access Journals (Sweden)

    Eman Ali Hussain

    2015-01-01

    Full Text Available Absract In this project A new method for solving Stochastic Differential Equations SDEs deriving by Wiener process numerically will be construct and implement using Accelerated Genetic Algorithm AGA. An SDE is a differential equation in which one or more of the terms and hence the solutions itself is a stochastic process. Solving stochastic differential equations requires going away from the recognizable deterministic setting of ordinary and partial differential equations into a world where the evolution of a quantity has an inherent random component and where the expected behavior of this quantity can be described in terms of probability distributions. We applied our method on the Ito formula which is equivalent to the SDE to find approximation solution of the SDEs. Numerical experiments illustrate the behavior of the proposed method.

  19. Factor Analysis with EM Algorithm Never Gives Improper Solutions when Sample Covariance and Initial Parameter Matrices Are Proper

    Science.gov (United States)

    Adachi, Kohei

    2013-01-01

    Rubin and Thayer ("Psychometrika," 47:69-76, 1982) proposed the EM algorithm for exploratory and confirmatory maximum likelihood factor analysis. In this paper, we prove the following fact: the EM algorithm always gives a proper solution with positive unique variances and factor correlations with absolute values that do not exceed one,…

  20. Smooth solutions for the dyadic model

    International Nuclear Information System (INIS)

    Barbato, David; Morandin, Francesco; Romito, Marco

    2011-01-01

    We consider the dyadic model, which is a toy model to test issues of well-posedness and blow-up for the Navier–Stokes and Euler equations. We prove well-posedness of positive solutions of the viscous problem in the relevant scaling range which corresponds to Navier–Stokes. Likewise we prove well-posedness for the inviscid problem (in a suitable regularity class) when the parameter corresponds to the strongest transport effect of the nonlinearity

  1. Survey of chemically amplified resist models and simulator algorithms

    Science.gov (United States)

    Croffie, Ebo H.; Yuan, Lei; Cheng, Mosong; Neureuther, Andrew R.

    2001-08-01

    Modeling has become indespensable tool for chemically amplified resist (CAR) evaluations. It has been used extensively to study acid diffusion and its effects on resist image formation. Several commercial and academic simulators have been developed for CAR process simulation. For commercial simulators such as PROLITH (Finle Technologies) and Solid-C (Sigma-C), the user is allowed to choose between an empirical model or a concentration dependant diffusion model. The empirical model is faster but not very accurate for 2-dimension resist simulations. In this case there is a trade off between the speed of the simulator and the accuracy of the results. An academic simulator such as STORM (U.C. Berkeley) gives the user a choice of different algorithms including Fast Imaging 2nd order finite difference algorithm and Moving Boundary finite element algorithm. A user interested in simulating the volume shrinkage and polymer stress effects during post exposure bake will need the Moving Boundary algorithm whereas a user interested in the latent image formation without polymer deformations will find the Fast Imaging algorithm more appropriate. The Fast Imaging algorithm is generally faster and requires less computer memory. This choice of algorithm presents a trade off between speed and level of detail in resist profile prediction. This paper surveys the different models and simulator algorithms available in the literature. Contributions in the field of CAR modeling including contributions to characterization of CAR exposure and post exposure bake (PEB) processes for different resist systems. Several numerical algorithms and their performances will also be discussed in this paper.

  2. A Developed Artificial Bee Colony Algorithm Based on Cloud Model

    Directory of Open Access Journals (Sweden)

    Ye Jin

    2018-04-01

    Full Text Available The Artificial Bee Colony (ABC algorithm is a bionic intelligent optimization method. The cloud model is a kind of uncertainty conversion model between a qualitative concept T ˜ that is presented by nature language and its quantitative expression, which integrates probability theory and the fuzzy mathematics. A developed ABC algorithm based on cloud model is proposed to enhance accuracy of the basic ABC algorithm and avoid getting trapped into local optima by introducing a new select mechanism, replacing the onlooker bees’ search formula and changing the scout bees’ updating formula. Experiments on CEC15 show that the new algorithm has a faster convergence speed and higher accuracy than the basic ABC and some cloud model based ABC variants.

  3. Comparison of parameter estimation algorithms in hydrological modelling

    DEFF Research Database (Denmark)

    Blasone, Roberta-Serena; Madsen, Henrik; Rosbjerg, Dan

    2006-01-01

    Local search methods have been applied successfully in calibration of simple groundwater models, but might fail in locating the optimum for models of increased complexity, due to the more complex shape of the response surface. Global search algorithms have been demonstrated to perform well...... for these types of models, although at a more expensive computational cost. The main purpose of this study is to investigate the performance of a global and a local parameter optimization algorithm, respectively, the Shuffled Complex Evolution (SCE) algorithm and the gradient-based Gauss......-Marquardt-Levenberg algorithm (implemented in the PEST software), when applied to a steady-state and a transient groundwater model. The results show that PEST can have severe problems in locating the global optimum and in being trapped in local regions of attractions. The global SCE procedure is, in general, more effective...

  4. Testing algorithms for a passenger train braking performance model.

    Science.gov (United States)

    2011-09-01

    "The Federal Railroad Administrations Office of Research and Development funded a project to establish performance model to develop, analyze, and test positive train control (PTC) braking algorithms for passenger train operations. With a good brak...

  5. Hybrid genetic algorithms: solutions in realistic dynamic and setup dependent job-shop scheduling problems

    Directory of Open Access Journals (Sweden)

    Rogério M. Branco

    2016-07-01

    Full Text Available This paper discusses the application of heuristic-based evolutionary technique in search for solutions concerning the dynamic job-shop scheduling problems with dependent setup times and alternate routes. With a combinatorial nature, these problems belong to an NP-hard class, with an aggravated condition when in realistic, dynamic and therefore, more complex cases than the traditional static ones. The proposed genetic algorithm executes two important functions: choose the routes using dispatching rules when forming each individual from a defined set of available machines and, also make the scheduling for each of these individuals created. The chromosome codifies a route, or the selected machines, and also an order to process the operations. In essence , each individual needs to be decoded by the scheduler to evaluate its time of completion, so the fitness function of the genetic algorithm, applying the modified Giffler and Thomson’s algorithm, obtains a scheduling of the selected routes in a given planning horizon. The scheduler considers the preparation time between operations on the machines and can manage operations exchange respecting the route and the order given by the chromosome. The best results in the evolutionary process are individuals with routes and processing orders optimized for this type of problema.

  6. Optimization of the K-means algorithm for the solution of high dimensional instances

    Science.gov (United States)

    Pérez, Joaquín; Pazos, Rodolfo; Olivares, Víctor; Hidalgo, Miguel; Ruiz, Jorge; Martínez, Alicia; Almanza, Nelva; González, Moisés

    2016-06-01

    This paper addresses the problem of clustering instances with a high number of dimensions. In particular, a new heuristic for reducing the complexity of the K-means algorithm is proposed. Traditionally, there are two approaches that deal with the clustering of instances with high dimensionality. The first executes a preprocessing step to remove those attributes of limited importance. The second, called divide and conquer, creates subsets that are clustered separately and later their results are integrated through post-processing. In contrast, this paper proposes a new solution which consists of the reduction of distance calculations from the objects to the centroids at the classification step. This heuristic is derived from the visual observation of the clustering process of K-means, in which it was found that the objects can only migrate to adjacent clusters without crossing distant clusters. Therefore, this heuristic can significantly reduce the number of distance calculations from an object to the centroids of the potential clusters that it may be classified to. To validate the proposed heuristic, it was designed a set of experiments with synthetic and high dimensional instances. One of the most notable results was obtained for an instance of 25,000 objects and 200 dimensions, where its execution time was reduced up to 96.5% and the quality of the solution decreased by only 0.24% when compared to the K-means algorithm.

  7. A Dynamic Traffic Signal Timing Model and its Algorithm for Junction of Urban Road

    DEFF Research Database (Denmark)

    Cai, Yanguang; Cai, Hao

    2012-01-01

    As an important part of Intelligent Transportation System, the scientific traffic signal timing of junction can improve the efficiency of urban transport. This paper presents a novel dynamic traffic signal timing model. According to the characteristics of the model, hybrid chaotic quantum......-time and dynamic signal control of junction. To obtain the optimal solution of the model by hybrid chaotic quantum evolutionary algorithm, the model is converted to an easily solvable form. To simplify calculation, we give the expression of the partial derivative and change rate of the objective function...... such that the implementation of the algorithm only involves function assignments and arithmetic operations and thus avoids complex operations such as integral and differential. Simulation results show that the algorithm has less remain vehicles than Webster method, higher convergence rate and convergence speed than quantum...

  8. Coarse-Grained Modeling of Polyelectrolyte Solutions

    Science.gov (United States)

    Denton, Alan R.; May, Sylvio

    2014-03-01

    Ionic mixtures, such as electrolyte and polyelectrolyte solutions, have attracted much attention recently for their rich and challenging combination of electrostatic and non-electrostatic interparticle forces and their practical importance, from battery technologies to biological systems. Hydration of ions in aqueous solutions is known to entail ion-specific effects, including variable solubility of organic molecules, as manifested in the classic Hofmeister series for salting-in and salting-out of proteins. The physical mechanism by which the solvent (water) mediates effective interactions between ions, however, is still poorly understood. Starting from a microscopic model of a polyelectrolyte solution, we apply a perturbation theory to derive a coarse-grained model of ions interacting through both long-range electrostatic and short-range solvent-induced pair potentials. Taking these effective interactions as input to molecular dynamics simulations, we calculate structural and thermodynamic properties of aqueous ionic solutions. This work was supported by the National Science Foundation under Grant No. DMR-1106331.

  9. An ensemble based nonlinear orthogonal matching pursuit algorithm for sparse history matching of reservoir models

    KAUST Repository

    Fsheikh, Ahmed H.

    2013-01-01

    A nonlinear orthogonal matching pursuit (NOMP) for sparse calibration of reservoir models is presented. Sparse calibration is a challenging problem as the unknowns are both the non-zero components of the solution and their associated weights. NOMP is a greedy algorithm that discovers at each iteration the most correlated components of the basis functions with the residual. The discovered basis (aka support) is augmented across the nonlinear iterations. Once the basis functions are selected from the dictionary, the solution is obtained by applying Tikhonov regularization. The proposed algorithm relies on approximate gradient estimation using an iterative stochastic ensemble method (ISEM). ISEM utilizes an ensemble of directional derivatives to efficiently approximate gradients. In the current study, the search space is parameterized using an overcomplete dictionary of basis functions built using the K-SVD algorithm.

  10. Near infrared system coupled chemometric algorithms for enumeration of total fungi count in cocoa beans neat solution.

    Science.gov (United States)

    Kutsanedzie, Felix Y H; Chen, Quansheng; Hassan, Md Mehedi; Yang, Mingxiu; Sun, Hao; Rahman, Md Hafizur

    2018-02-01

    Total fungi count (TFC) is a quality indicator of cocoa beans when unmonitored leads to quality and safety problems. Fourier transform near infrared spectroscopy (FT-NIRS) combined with chemometric algorithms like partial least square (PLS); synergy interval-PLS (Si-PLS); synergy interval-genetic algorithm-PLS (Si-GAPLS); Ant colony optimization - PLS (ACO-PLS) and competitive-adaptive reweighted sampling-PLS (CARS-PLS) was employed to predict TFC in cocoa beans neat solution. Model results were evaluated using the correlation coefficients of the prediction (Rp) and calibration (Rc); root mean square error of prediction (RMSEP), and the ratio of sample standard deviation to RMSEP (RPD). The developed models performance yielded 0.951≤Rp≤0.975; and 3.15≤RPD≤4.32. The models' prediction stability improved in the order of PLS

  11. Algorithms

    Indian Academy of Sciences (India)

    algorithms such as synthetic (polynomial) division have been found in Vedic Mathematics which are dated much before Euclid's algorithm. A programming language ... ·1 x:=sln(theta) x : = sm(theta) 1. ~. Idl d.t Read A.B,C. ~ lei ~ Print x.y.z. L;;;J. Figure 2 Symbols used In flowchart language to rep- resent Assignment, Read.

  12. Algorithms

    Indian Academy of Sciences (India)

    In the previous articles, we have discussed various common data-structures such as arrays, lists, queues and trees and illustrated the widely used algorithm design paradigm referred to as 'divide-and-conquer'. Although there has been a large effort in realizing efficient algorithms, there are not many universally accepted ...

  13. Numerical solution of a model for a superconductor field problem

    International Nuclear Information System (INIS)

    Alsop, L.E.; Goodman, A.S.; Gustavson, F.G.; Miranker, W.L.

    1979-01-01

    A model of a magnetic field problem occurring in connection with Josephson junction devices is derived, and numerical solutions are obtained. The model is of mathematical interest, because the magnetic vector potential satisfies inhomogeneous Helmholtz equations in part of the region, i.e., the superconductors, and the Laplace equation elsewhere. Moreover, the inhomogeneities are the guage constants for the potential, which are different for each superconductor, and their magnitudes are proportional to the currents flowing in the superconductors. These constants are directly related to the self and mutual inductances of the superconducting elements in the device. The numerical solution is obtained by the iterative use of a fast Poisson solver. Chebyshev acceleration is used to reduce the number of iterations required to obtain a solution. A typical problem involves solving 100,000 simultaneous equations, which the algorithm used with this model does in 20 iterations, requiring three minutes of CPU time on an IBM VM/370/168. Excellent agreement is obtained between calculated and observed values for the inductances

  14. Drexel University Shell Model (DUSM) algorithm

    Science.gov (United States)

    Valliéres, Michel; Novoselsky, Akiva

    1994-03-01

    This lecture is devoted to the Drexel University Shell Model (DUSM) code; this is a new shell-model code based on a separation of the various subspaces in which the single particle wavefunctions are defined. This is achieved via extensive use of permutation group concepts and a redefinition of the Coeficients of Fractional Parentage (CFP) to include permutation labels. This leads to a modern and efficient approach to nuclear shell-model.

  15. Drexel University Shell Model (DUSM) algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Vallieres, M. (Drexel Univ., Philadelphia, PA (United States). Dept. of Physics and Atmospheric Science); Novoselsky, A. (Hebrew Univ., Jerusalem (Israel). Dept. of Physics)

    1994-03-28

    This lecture is devoted to the Drexel University Shell Model (DUSM) code; this is a new shell-model code based on a separation of the various subspaces in which the single particle wavefunctions are defined. This is achieved via extensive use of permutation group concepts and a redefinition of the Coeficients of Fractional Parentage (CEP) to include permutation labels. This leads to a modern and efficient approach to nuclear shell-model. (orig.)

  16. Gas Emission Prediction Model of Coal Mine Based on CSBP Algorithm

    Directory of Open Access Journals (Sweden)

    Xiong Yan

    2016-01-01

    Full Text Available In view of the nonlinear characteristics of gas emission in a coal working face, a prediction method is proposed based on cuckoo search algorithm optimized BP neural network (CSBP. In the CSBP algorithm, the cuckoo search is adopted to optimize weight and threshold parameters of BP network, and obtains the global optimal solutions. Furthermore, the twelve main affecting factors of the gas emission in the coal working face are taken as input vectors of CSBP algorithm, the gas emission is acted as output vector, and then the prediction model of BP neural network with optimal parameters is established. The results show that the CSBP algorithm has batter generalization ability and higher prediction accuracy, and can be utilized effectively in the prediction of coal mine gas emission.

  17. Improved CHAID algorithm for document structure modelling

    Science.gov (United States)

    Belaïd, A.; Moinel, T.; Rangoni, Y.

    2010-01-01

    This paper proposes a technique for the logical labelling of document images. It makes use of a decision-tree based approach to learn and then recognise the logical elements of a page. A state-of-the-art OCR gives the physical features needed by the system. Each block of text is extracted during the layout analysis and raw physical features are collected and stored in the ALTO format. The data-mining method employed here is the "Improved CHi-squared Automatic Interaction Detection" (I-CHAID). The contribution of this work is the insertion of logical rules extracted from the logical layout knowledge to support the decision tree. Two setups have been tested; the first uses one tree per logical element, the second one uses a single tree for all the logical elements we want to recognise. The main system, implemented in Java, coordinates the third-party tools (Omnipage for the OCR part, and SIPINA for the I-CHAID algorithm) using XML and XSL transforms. It was tested on around 1000 documents belonging to the ICPR'04 and ICPR'08 conference proceedings, representing about 16,000 blocks. The final error rate for determining the logical labels (among 9 different ones) is less than 6%.

  18. Integrating R and Java for Enhancing Interactivity of Algorithmic Data Analysis Software Solutions

    Directory of Open Access Journals (Sweden)

    Titus Felix FURTUNĂ

    2016-06-01

    Full Text Available Conceiving software solutions for statistical processing and algorithmic data analysis involves handling diverse data, fetched from various sources and in different formats, and presenting the results in a suggestive, tailorable manner. Our ongoing research aims to design programming technics for integrating R developing environment with Java programming language for interoperability at a source code level. The goal is to combine the intensive data processing capabilities of R programing language, along with the multitude of statistical function libraries, with the flexibility offered by Java programming language and platform, in terms of graphical user interface and mathematical function libraries. Both developing environments are multiplatform oriented, and can complement each other through interoperability. R is a comprehensive and concise programming language, benefiting from a continuously expanding and evolving set of packages for statistical analysis, developed by the open source community. While is a very efficient environment for statistical data processing, R platform lacks support for developing user friendly, interactive, graphical user interfaces (GUIs. Java on the other hand, is a high level object oriented programming language, which supports designing and developing performant and interactive frameworks for general purpose software solutions, through Java Foundation Classes, JavaFX and various graphical libraries. In this paper we treat both aspects of integration and interoperability that refer to integrating Java code into R applications, and bringing R processing sequences into Java driven software solutions. Our research has been conducted focusing on case studies concerning pattern recognition and cluster analysis.

  19. Solution of the optimal plant location and sizing problem using simulated annealing and genetic algorithms

    Energy Technology Data Exchange (ETDEWEB)

    Rao, R.; Buescher, K.L.; Hanagandi, V.

    1995-12-31

    In the optimal plant location and sizing problem it is desired to optimize cost function involving plant sizes, locations, and production schedules in the face of supply-demand and plant capacity constraints. We will use simulated annealing (SA) and a genetic algorithm (GA) to solve this problem. We will compare these techniques with respect to computational expenses, constraint handling capabilities, and the quality of the solution obtained in general. Simulated Annealing is a combinatorial stochastic optimization technique which has been shown to be effective in obtaining fast suboptimal solutions for computationally, hard problems. The technique is especially attractive since solutions are obtained in polynomial time for problems where an exhaustive search for the global optimum would require exponential time. We propose a synergy between the cluster analysis technique, popular in classical stochastic global optimization, and the GA to accomplish global optimization. This synergy minimizes redundant searches around local optima and enhances the capable it of the GA to explore new areas in the search space.

  20. Immune System Model Calibration by Genetic Algorithm

    NARCIS (Netherlands)

    Presbitero, A.; Krzhizhanovskaya, V.; Mancini, E.; Brands, R.; Sloot, P.

    2016-01-01

    We aim to develop a mathematical model of the human immune system for advanced individualized healthcare where medication plan is fine-tuned to fit a patient's conditions through monitored biochemical processes. One of the challenges is calibrating model parameters to satisfy existing experimental

  1. Approximation Algorithms for Model-Based Diagnosis

    NARCIS (Netherlands)

    Feldman, A.B.

    2010-01-01

    Model-based diagnosis is an area of abductive inference that uses a system model, together with observations about system behavior, to isolate sets of faulty components (diagnoses) that explain the observed behavior, according to some minimality criterion. This thesis presents greedy approximation

  2. Stochastic cluster algorithms for discrete Gaussian (SOS) models

    International Nuclear Information System (INIS)

    Evertz, H.G.; Hamburg Univ.; Hasenbusch, M.; Marcu, M.; Tel Aviv Univ.; Pinn, K.; Muenster Univ.; Solomon, S.

    1990-10-01

    We present new Monte Carlo cluster algorithms which eliminate critical slowing down in the simulation of solid-on-solid models. In this letter we focus on the two-dimensional discrete Gaussian model. The algorithms are based on reflecting the integer valued spin variables with respect to appropriately chosen reflection planes. The proper choice of the reflection plane turns out to be crucial in order to obtain a small dynamical exponent z. Actually, the successful versions of our algorithm are a mixture of two different procedures for choosing the reflection plane, one of them ergodic but slow, the other one non-ergodic and also slow when combined with a Metropolis algorithm. (orig.)

  3. Algebraic solution of the Stein-Stein model for stochastic volatility

    Science.gov (United States)

    Sophocleous, C.; O'Hara, J. G.; Leach, P. G. L.

    2011-04-01

    We provide an algebraic approach to the solution of the Stein-Stein model for stochastic volatility which arises in the determination of the Radon-Nikodym density of the minimal entropy of the martingale measure. We extend our investigation to the case in which the parameters of the model are time-dependent. Our algorithmic approach obviates the need for Ansätze for the structure of the solution.

  4. An evaporation model of multicomponent solution drops

    Science.gov (United States)

    Sartori, Silvana; Liñán, Amable; Lasheras, Juan C.

    2010-11-01

    Solutions of polymers are widely used in the pharmaceutical industry as tablets coatings. These allow controlling the rate at which the drug is delivered, taste or appearance. The coating is performed by spraying and drying the tablets at moderate temperatures. The wetting of the coating solution on the pill's surface depends on the droplet Webber and Re numbers, angle of impact and on the rheological properties of the droplet. We present a model for the evaporation of multicomponent solutions droplets in a hot air environment with temperatures substantially lower than the boiling temperature of the solvent. As the liquid vaporizes from the surface the fluid in the drop increases in concentration, until reaching its saturation point. After saturation, precipitation occurs uniformly within the drop. As the surface regresses, a compacting front formed by the precipitate at its maximum packing density advances into the drop, while the solute continues precipitating uniformly. This porous shell grows fast due to the double effect of surface regression and precipitation. The evaporation rate is determined by the rates at which heat is transported to the droplet surface and at which liquid vapor diffuses away from it. When the drop is fully compacted, the evaporation is drastically reduced.

  5. Applications of Flocking Algorithms to Input Modeling for Agent Movement

    Science.gov (United States)

    2011-12-01

    2445 Singham, Therkildsen, and Schruben We apply the following flocking algorithm to this leading boid to generate followers, who will then be mapped...due to the paths crossing. 2447 Singham, Therkildsen, and Schruben Figure 2: Plot of the path of a boid generated by the Group 4 flocking algorithm ...on the possible inputs. This method uses techniques from agent-based modeling to generate a flock of boids that follow the data. In this paper, we

  6. SU-F-R-10: Selecting the Optimal Solution for Multi-Objective Radiomics Model

    International Nuclear Information System (INIS)

    Zhou, Z; Folkert, M; Wang, J

    2016-01-01

    Purpose: To develop an evidential reasoning approach for selecting the optimal solution from a Pareto solution set obtained by a multi-objective radiomics model for predicting distant failure in lung SBRT. Methods: In the multi-objective radiomics model, both sensitivity and specificity are considered as the objective functions simultaneously. A Pareto solution set with many feasible solutions will be resulted from the multi-objective optimization. In this work, an optimal solution Selection methodology for Multi-Objective radiomics Learning model using the Evidential Reasoning approach (SMOLER) was proposed to select the optimal solution from the Pareto solution set. The proposed SMOLER method used the evidential reasoning approach to calculate the utility of each solution based on pre-set optimal solution selection rules. The solution with the highest utility was chosen as the optimal solution. In SMOLER, an optimal learning model coupled with clonal selection algorithm was used to optimize model parameters. In this study, PET, CT image features and clinical parameters were utilized for predicting distant failure in lung SBRT. Results: Total 126 solution sets were generated by adjusting predictive model parameters. Each Pareto set contains 100 feasible solutions. The solution selected by SMOLER within each Pareto set was compared to the manually selected optimal solution. Five-cross-validation was used to evaluate the optimal solution selection accuracy of SMOLER. The selection accuracies for five folds were 80.00%, 69.23%, 84.00%, 84.00%, 80.00%, respectively. Conclusion: An optimal solution selection methodology for multi-objective radiomics learning model using the evidential reasoning approach (SMOLER) was proposed. Experimental results show that the optimal solution can be found in approximately 80% cases.

  7. Solution of wind integrated thermal generation system for environmental optimal power flow using hybrid algorithm

    Directory of Open Access Journals (Sweden)

    Ambarish Panda

    2016-09-01

    Full Text Available A new evolutionary hybrid algorithm (HA has been proposed in this work for environmental optimal power flow (EOPF problem. The EOPF problem has been formulated in a nonlinear constrained multi objective optimization framework. Considering the intermittency of available wind power a cost model of the wind and thermal generation system is developed. Suitably formed objective function considering the operational cost, cost of emission, real power loss and cost of installation of FACTS devices for maintaining a stable voltage in the system has been optimized with HA and compared with particle swarm optimization algorithm (PSOA to prove its effectiveness. All the simulations are carried out in MATLAB/SIMULINK environment taking IEEE30 bus as the test system.

  8. An Algorithm for Optimally Fitting a Wiener Model

    Directory of Open Access Journals (Sweden)

    Lucas P. Beverlin

    2011-01-01

    Full Text Available The purpose of this work is to present a new methodology for fitting Wiener networks to datasets with a large number of variables. Wiener networks have the ability to model a wide range of data types, and their structures can yield parameters with phenomenological meaning. There are several challenges to fitting such a model: model stiffness, the nonlinear nature of a Wiener network, possible overfitting, and the large number of parameters inherent with large input sets. This work describes a methodology to overcome these challenges by using several iterative algorithms under supervised learning and fitting subsets of the parameters at a time. This methodology is applied to Wiener networks that are used to predict blood glucose concentrations. The predictions of validation sets from models fit to four subjects using this methodology yielded a higher correlation between observed and predicted observations than other algorithms, including the Gauss-Newton and Levenberg-Marquardt algorithms.

  9. A finite state projection algorithm for the stationary solution of the chemical master equation

    Science.gov (United States)

    Gupta, Ankit; Mikelson, Jan; Khammash, Mustafa

    2017-10-01

    The chemical master equation (CME) is frequently used in systems biology to quantify the effects of stochastic fluctuations that arise due to biomolecular species with low copy numbers. The CME is a system of ordinary differential equations that describes the evolution of probability density for each population vector in the state-space of the stochastic reaction dynamics. For many examples of interest, this state-space is infinite, making it difficult to obtain exact solutions of the CME. To deal with this problem, the Finite State Projection (FSP) algorithm was developed by Munsky and Khammash [J. Chem. Phys. 124(4), 044104 (2006)], to provide approximate solutions to the CME by truncating the state-space. The FSP works well for finite time-periods but it cannot be used for estimating the stationary solutions of CMEs, which are often of interest in systems biology. The aim of this paper is to develop a version of FSP which we refer to as the stationary FSP (sFSP) that allows one to obtain accurate approximations of the stationary solutions of a CME by solving a finite linear-algebraic system that yields the stationary distribution of a continuous-time Markov chain over the truncated state-space. We derive bounds for the approximation error incurred by sFSP and we establish that under certain stability conditions, these errors can be made arbitrarily small by appropriately expanding the truncated state-space. We provide several examples to illustrate our sFSP method and demonstrate its efficiency in estimating the stationary distributions. In particular, we show that using a quantized tensor-train implementation of our sFSP method, problems admitting more than 100 × 106 states can be efficiently solved.

  10. Co-clustering models, algorithms and applications

    CERN Document Server

    Govaert, Gérard

    2013-01-01

    Cluster or co-cluster analyses are important tools in a variety of scientific areas. The introduction of this book presents a state of the art of already well-established, as well as more recent methods of co-clustering. The authors mainly deal with the two-mode partitioning under different approaches, but pay particular attention to a probabilistic approach. Chapter 1 concerns clustering in general and the model-based clustering in particular. The authors briefly review the classical clustering methods and focus on the mixture model. They present and discuss the use of different mixture

  11. Decentralized Fuzzy P-hub Centre Problem: Extended Model and Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    Sara Mousavinia

    2017-02-01

    Full Text Available This paper studies the uncapacitated P-hub center problem in a network under decentralized management assuming time as a fuzzy variable. In this network, transport companies act independently, each company makes its route choices according to its own criteria. In this model, time is presented by triangular fuzzy number and used to calculate the fraction of users that probably choose hub routes instead of direct routes. To solve the problem, two genetic algorithms are proposed. The computational results compared with LINGO indicate that the proposed algorithm solves large-scale instances within promising computational time and outperforms LINGO in terms of solution quality.

  12. Economic Models and Algorithms for Distributed Systems

    CERN Document Server

    Neumann, Dirk; Altmann, Jorn; Rana, Omer F

    2009-01-01

    Distributed computing models for sharing resources such as Grids, Peer-to-Peer systems, or voluntary computing are becoming increasingly popular. This book intends to discover fresh avenues of research and amendments to existing technologies, aiming at the successful deployment of commercial distributed systems

  13. Robust Return Algorithm for Anisotropic Plasticity Models

    DEFF Research Database (Denmark)

    Tidemann, L.; Krenk, Steen

    2017-01-01

    Plasticity models can be defined by an energy potential, a plastic flow potential and a yield surface. The energy potential defines the relation between the observable elastic strains ϒe and the energy conjugate stresses Τe and between the non-observable internal strains i and the energy conjugat...

  14. An Integrated Multi-Echelon Supply Chain Network Design Considering Stochastic Demand: A Genetic Algorithm Based Solution

    Directory of Open Access Journals (Sweden)

    Sara Nakhjirkan

    2017-09-01

    in green supply chain. Vehicle routing between distribution centres and customers has been considered in the model. Establishment place of distribution centres among potential places is determined by the model. The distributors use continuous review policy (r, Q to control the inventory. The proposed model object is to find an optimal supply chain with minimum costs. To validate the proposed model and measure its compliance with real world problems, GAMS IDE/Cplex has been used. In order to measure the efficiency of the proposed model in large scale problems, a genetic algorithm has been used. The results confirm the efficiency of the proposed model as a practical tool for decision makers to solve location-inventory-routing problems in green supply chain. The proposed GA could reduce the solving time by 85% while reaching on the average 97% of optimal solution compared with exact method.

  15. Data mining concepts models methods and algorithms

    CERN Document Server

    Kantardzic, Mehmed

    2011-01-01

    This book reviews state-of-the-art methodologies and techniques for analyzing enormous quantities of raw data in high-dimensional data spaces, to extract new information for decision making. The goal of this book is to provide a single introductory source, organized in a systematic way, in which we could direct the readers in analysis of large data sets, through the explanation of basic concepts, models and methodologies developed in recent decades.

  16. A Formal Verification Model for Performance Analysis of Reinforcement Learning Algorithms Applied t o Dynamic Networks

    Directory of Open Access Journals (Sweden)

    Shrirang Ambaji KULKARNI

    2017-04-01

    Full Text Available Routing data packets in a dynamic network is a difficult and important problem in computer networks. As the network is dynamic, it is subject to frequent topology changes and is subject to variable link costs due to congestion and bandwidth. Existing shortest path algorithms fail to converge to better solutions under dynamic network conditions. Reinforcement learning algorithms posses better adaptation techniques in dynamic environments. In this paper we apply model based Q-Routing technique for routing in dynamic network. To analyze the correctness of Q-Routing algorithms mathematically, we provide a proof and also implement a SPIN based verification model. We also perform simulation based analysis of Q-Routing for given metrics.

  17. Algorithms for Optimal Model Distributions in Adaptive Switching Control Schemes

    Directory of Open Access Journals (Sweden)

    Debarghya Ghosh

    2016-03-01

    Full Text Available Several multiple model adaptive control architectures have been proposed in the literature. Despite many advances in theory, the crucial question of how to synthesize the pairs model/controller in a structurally optimal way is to a large extent not addressed. In particular, it is not clear how to place the pairs model/controller is such a way that the properties of the switching algorithm (e.g., number of switches, learning transient, final performance are optimal with respect to some criteria. In this work, we focus on the so-called multi-model unfalsified adaptive supervisory switching control (MUASSC scheme; we define a suitable structural optimality criterion and develop algorithms for synthesizing the pairs model/controller in such a way that they are optimal with respect to the structural optimality criterion we defined. The peculiarity of the proposed optimality criterion and algorithms is that the optimization is carried out so as to optimize the entire behavior of the adaptive algorithm, i.e., both the learning transient and the steady-state response. A comparison is made with respect to the model distribution of the robust multiple model adaptive control (RMMAC, where the optimization considers only the steady-state ideal response and neglects any learning transient.

  18. Development of Gis Tool for the Solution of Minimum Spanning Tree Problem using Prim's Algorithm

    Science.gov (United States)

    Dutta, S.; Patra, D.; Shankar, H.; Alok Verma, P.

    2014-11-01

    minimum spanning tree (MST) of a connected, undirected and weighted network is a tree of that network consisting of all its nodes and the sum of weights of all its edges is minimum among all such possible spanning trees of the same network. In this study, we have developed a new GIS tool using most commonly known rudimentary algorithm called Prim's algorithm to construct the minimum spanning tree of a connected, undirected and weighted road network. This algorithm is based on the weight (adjacency) matrix of a weighted network and helps to solve complex network MST problem easily, efficiently and effectively. The selection of the appropriate algorithm is very essential otherwise it will be very hard to get an optimal result. In case of Road Transportation Network, it is very essential to find the optimal results by considering all the necessary points based on cost factor (time or distance). This paper is based on solving the Minimum Spanning Tree (MST) problem of a road network by finding it's minimum span by considering all the important network junction point. GIS technology is usually used to solve the network related problems like the optimal path problem, travelling salesman problem, vehicle routing problems, location-allocation problems etc. Therefore, in this study we have developed a customized GIS tool using Python script in ArcGIS software for the solution of MST problem for a Road Transportation Network of Dehradun city by considering distance and time as the impedance (cost) factors. It has a number of advantages like the users do not need a greater knowledge of the subject as the tool is user-friendly and that allows to access information varied and adapted the needs of the users. This GIS tool for MST can be applied for a nationwide plan called Prime Minister Gram Sadak Yojana in India to provide optimal all weather road connectivity to unconnected villages (points). This tool is also useful for constructing highways or railways spanning several

  19. A tuning algorithm for model predictive controllers based on genetic algorithms and fuzzy decision making.

    Science.gov (United States)

    van der Lee, J H; Svrcek, W Y; Young, B R

    2008-01-01

    Model Predictive Control is a valuable tool for the process control engineer in a wide variety of applications. Because of this the structure of an MPC can vary dramatically from application to application. There have been a number of works dedicated to MPC tuning for specific cases. Since MPCs can differ significantly, this means that these tuning methods become inapplicable and a trial and error tuning approach must be used. This can be quite time consuming and can result in non-optimum tuning. In an attempt to resolve this, a generalized automated tuning algorithm for MPCs was developed. This approach is numerically based and combines a genetic algorithm with multi-objective fuzzy decision-making. The key advantages to this approach are that genetic algorithms are not problem specific and only need to be adapted to account for the number and ranges of tuning parameters for a given MPC. As well, multi-objective fuzzy decision-making can handle qualitative statements of what optimum control is, in addition to being able to use multiple inputs to determine tuning parameters that best match the desired results. This is particularly useful for multi-input, multi-output (MIMO) cases where the definition of "optimum" control is subject to the opinion of the control engineer tuning the system. A case study will be presented in order to illustrate the use of the tuning algorithm. This will include how different definitions of "optimum" control can arise, and how they are accounted for in the multi-objective decision making algorithm. The resulting tuning parameters from each of the definition sets will be compared, and in doing so show that the tuning parameters vary in order to meet each definition of optimum control, thus showing the generalized automated tuning algorithm approach for tuning MPCs is feasible.

  20. Research on the time optimization model algorithm of Customer Collaborative Product Innovation

    Directory of Open Access Journals (Sweden)

    Guodong Yu

    2014-01-01

    Full Text Available Purpose: To improve the efficiency of information sharing among the innovation agents of customer collaborative product innovation and shorten the product design cycle, an improved genetic annealing algorithm of the time optimization was presented. Design/methodology/approach: Based on the analysis of the objective relationship between the design tasks, the paper takes job shop problems for machining model and proposes the improved genetic algorithm to solve the problems, which is based on the niche technology and thus a better product collaborative innovation design time schedule is got to improve the efficiency. Finally, through the collaborative innovation design of a certain type of mobile phone, the proposed model and method were verified to be correct and effective. Findings and Originality/value: An algorithm with obvious advantages in terms of searching capability and optimization efficiency of customer collaborative product innovation was proposed. According to the defects of the traditional genetic annealing algorithm, the niche genetic annealing algorithm was presented. Firstly, it avoided the effective gene deletions at the early search stage and guaranteed the diversity of solution; Secondly, adaptive double point crossover and swap mutation strategy were introduced to overcome the defects of long solving process and easily converging local minimum value due to the fixed crossover and mutation probability; Thirdly, elite reserved strategy was imported that optimal solution missing was avoided effectively and evolution speed was accelerated. Originality/value: Firstly, the improved genetic simulated annealing algorithm overcomes some defects such as effective gene easily lost in early search. It is helpful to shorten the calculation process and improve the accuracy of the convergence value. Moreover, it speeds up the evolution and ensures the reliability of the optimal solution. Meanwhile, it has obvious advantages in efficiency of

  1. The use of multigrid techniques in the solution of the Elrod algorithm for a dynamically loaded journal bearing. M.S. Thesis. Final Report

    Science.gov (United States)

    Woods, Claudia M.

    1988-01-01

    A numerical solution to a theoretical model of vapor cavitation in a dynamically loaded journal bearing is developed, utilizing a multigrid iterative technique. The code is compared with a presently existing direct solution in terms of computational time and accuracy. The model is based on the Elrod algorithm, a control volume approach to the Reynolds equation which mimics the Jakobssen-Floberg and Olsson cavitation theory. Besides accounting for a moving cavitation boundary and conservation of mass at the boundary, it also conserves mass within the cavitated region via liquid striations. The mixed nature of the equations (elliptic in the full film zone and nonelliptic in the cavitated zone) coupled with the dynamic aspects of the problem create interesting difficulties for the present solution approach. Emphasis is placed on the methods found to eliminate solution instabilities. Excellent results are obtained for both accuracy and reduction of computational time.

  2. Second-order p-iterative solution of the Lambert/Gauss problem. [algorithm for efficient orbit determination

    Science.gov (United States)

    Boltz, F. W.

    1984-01-01

    An algorithm is presented for efficient p-iterative solution of the Lambert/Gauss orbit-determination problem using second-order Newton iteration. The algorithm is based on a universal transformation of Kepler's time-of-flight equation and approximate inverse solutions of this equation for short-way and long-way flight paths. The approximate solutions provide both good starting values for iteration and simplified computation of the second-order term in the iteration formula. Numerical results are presented which indicate that in many cases of practical significance (except those having collinear position vectors) the algorithm produces at least eight significant digits of accuracy with just two or three steps of iteration.

  3. A genetic-algorithm-aided stochastic optimization model for regional air quality management under uncertainty.

    Science.gov (United States)

    Qin, Xiaosheng; Huang, Guohe; Liu, Lei

    2010-01-01

    A genetic-algorithm-aided stochastic optimization (GASO) model was developed in this study for supporting regional air quality management under uncertainty. The model incorporated genetic algorithm (GA) and Monte Carlo simulation techniques into a general stochastic chance-constrained programming (CCP) framework and allowed uncertainties in simulation and optimization model parameters to be considered explicitly in the design of least-cost strategies. GA was used to seek the optimal solution of the management model by progressively evaluating the performances of individual solutions. Monte Carlo simulation was used to check the feasibility of each solution. A management problem in terms of regional air pollution control was studied to demonstrate the applicability of the proposed method. Results of the case study indicated the proposed model could effectively communicate uncertainties into the optimization process and generate solutions that contained a spectrum of potential air pollutant treatment options with risk and cost information. Decision alternatives could be obtained by analyzing tradeoffs between the overall pollutant treatment cost and the system-failure risk due to inherent uncertainties.

  4. Modeling Algorithms in SystemC and ACL2

    Directory of Open Access Journals (Sweden)

    John W. O'Leary

    2014-06-01

    Full Text Available We describe the formal language MASC, based on a subset of SystemC and intended for modeling algorithms to be implemented in hardware. By means of a special-purpose parser, an algorithm coded in SystemC is converted to a MASC model for the purpose of documentation, which in turn is translated to ACL2 for formal verification. The parser also generates a SystemC variant that is suitable as input to a high-level synthesis tool. As an illustration of this methodology, we describe a proof of correctness of a simple 32-bit radix-4 multiplier.

  5. Methodology, models and algorithms in thermographic diagnostics

    CERN Document Server

    Živčák, Jozef; Madarász, Ladislav; Rudas, Imre J

    2013-01-01

    This book presents  the methodology and techniques of  thermographic applications with focus primarily on medical thermography implemented for parametrizing the diagnostics of the human body. The first part of the book describes the basics of infrared thermography, the possibilities of thermographic diagnostics and the physical nature of thermography. The second half includes tools of intelligent engineering applied for the solving of selected applications and projects. Thermographic diagnostics was applied to problematics of paraplegia and tetraplegia and carpal tunnel syndrome (CTS). The results of the research activities were created with the cooperation of the four projects within the Ministry of Education, Science, Research and Sport of the Slovak Republic entitled Digital control of complex systems with two degrees of freedom, Progressive methods of education in the area of control and modeling of complex object oriented systems on aircraft turbocompressor engines, Center for research of control of te...

  6. Introduction to genetic algorithms as a modeling tool

    International Nuclear Information System (INIS)

    Wildberger, A.M.; Hickok, K.A.

    1990-01-01

    Genetic algorithms are search and classification techniques modeled on natural adaptive systems. This is an introduction to their use as a modeling tool with emphasis on prospects for their application in the power industry. It is intended to provide enough background information for its audience to begin to follow technical developments in genetic algorithms and to recognize those which might impact on electric power engineering. Beginning with a discussion of genetic algorithms and their origin as a model of biological adaptation, their advantages and disadvantages are described in comparison with other modeling tools such as simulation and neural networks in order to provide guidance in selecting appropriate applications. In particular, their use is described for improving expert systems from actual data and they are suggested as an aid in building mathematical models. Using the Thermal Performance Advisor as an example, it is suggested how genetic algorithms might be used to make a conventional expert system and mathematical model of a power plant adapt automatically to changes in the plant's characteristics

  7. Modeling of temperatures by using the algorithm of queue burning movement in the UCG Process

    Directory of Open Access Journals (Sweden)

    Milan Durdán

    2015-10-01

    Full Text Available In this contribution, a proposal of the system for indirect measurement temperatures in the underground coal gasification (UCG process is presented. A two-dimensional solution results from the Fourier partial differential equation of the heat conduction was used for the calculation of the temperature field in the real coal seam. An algorithm of queue burning movement for modeling the boundary conditions in gasification channel was created. Indirect measurement temperatures system was verified in the laboratory conditions.

  8. How effective and efficient are multiobjective evolutionary algorithms at hydrologic model calibration?

    Directory of Open Access Journals (Sweden)

    Y. Tang

    2006-01-01

    Full Text Available This study provides a comprehensive assessment of state-of-the-art evolutionary multiobjective optimization (EMO tools' relative effectiveness in calibrating hydrologic models. The relative computational efficiency, accuracy, and ease-of-use of the following EMO algorithms are tested: Epsilon Dominance Nondominated Sorted Genetic Algorithm-II (ε-NSGAII, the Multiobjective Shuffled Complex Evolution Metropolis algorithm (MOSCEM-UA, and the Strength Pareto Evolutionary Algorithm 2 (SPEA2. This study uses three test cases to compare the algorithms' performances: (1 a standardized test function suite from the computer science literature, (2 a benchmark hydrologic calibration test case for the Leaf River near Collins, Mississippi, and (3 a computationally intensive integrated surface-subsurface model application in the Shale Hills watershed in Pennsylvania. One challenge and contribution of this work is the development of a methodology for comprehensively comparing EMO algorithms that have different search operators and randomization techniques. Overall, SPEA2 attained competitive to superior results for most of the problems tested in this study. The primary strengths of the SPEA2 algorithm lie in its search reliability and its diversity preservation operator. The biggest challenge in maximizing the performance of SPEA2 lies in specifying an effective archive size without a priori knowledge of the Pareto set. In practice, this would require significant trial-and-error analysis, which is problematic for more complex, computationally intensive calibration applications. ε-NSGAII appears to be superior to MOSCEM-UA and competitive with SPEA2 for hydrologic model calibration. ε-NSGAII's primary strength lies in its ease-of-use due to its dynamic population sizing and archiving which lead to rapid convergence to very high quality solutions with minimal user input. MOSCEM-UA is best suited for hydrologic model calibration applications that have small

  9. Algorithms

    Indian Academy of Sciences (India)

    like programming language. Recursion. One of the usual techniques of problem solving is to break the problem into smaller problems. From the solution of these smaller problems, one obtains a solution for the original problem. Consider the procedural abstraction described above. It is possible to visualize the given ...

  10. Uncertainty analysis of hydrological modeling in a tropical area using different algorithms

    Science.gov (United States)

    Rafiei Emam, Ammar; Kappas, Martin; Fassnacht, Steven; Linh, Nguyen Hoang Khanh

    2018-01-01

    Hydrological modeling outputs are subject to uncertainty resulting from different sources of errors (e.g., error in input data, model structure, and model parameters), making quantification of uncertainty in hydrological modeling imperative and meant to improve reliability of modeling results. The uncertainty analysis must solve difficulties in calibration of hydrological models, which further increase in areas with data scarcity. The purpose of this study is to apply four uncertainty analysis algorithms to a semi-distributed hydrological model, quantifying different source of uncertainties (especially parameter uncertainty) and evaluate their performance. In this study, the Soil and Water Assessment Tools (SWAT) eco-hydrological model was implemented for the watershed in the center of Vietnam. The sensitivity of parameters was analyzed, and the model was calibrated. The uncertainty analysis for the hydrological model was conducted based on four algorithms: Generalized Likelihood Uncertainty Estimation (GLUE), Sequential Uncertainty Fitting (SUFI), Parameter Solution method (ParaSol) and Particle Swarm Optimization (PSO). The performance of the algorithms was compared using P-factor and Rfactor, coefficient of determination (R 2), the Nash Sutcliffe coefficient of efficiency (NSE) and Percent Bias (PBIAS). The results showed the high performance of SUFI and PSO with P-factor>0.83, R-factor 0.91, NSE>0.89, and 0.18model use for policy or management decisions.

  11. Calibration of microscopic traffic simulation models using metaheuristic algorithms

    Directory of Open Access Journals (Sweden)

    Miao Yu

    2017-06-01

    Full Text Available This paper presents several metaheuristic algorithms to calibrate a microscopic traffic simulation model. The genetic algorithm (GA, Tabu Search (TS, and a combination of the GA and TS (i.e., warmed GA and warmed TS are implemented and compared. A set of traffic data collected from the I-5 Freeway, Los Angles, California, is used. Objective functions are defined to minimize the difference between simulated and field traffic data which are built based on the flow and speed. Several car-following parameters in VISSIM, which can significantly affect the simulation outputs, are selected to calibrate. A better match to the field measurements is reached with the GA, TS, and warmed GA and TS when comparing with that only using the default parameters in VISSIM. Overall, TS performs very well and can be used to calibrate parameters. Combining metaheuristic algorithms clearly performs better and therefore is highly recommended for calibrating microscopic traffic simulation models.

  12. Algorithms

    Indian Academy of Sciences (India)

    In the program shown in Figure 1, we have repeated the algorithm. M times and we can make the following observations. Each block is essentially a different instance of "code"; that is, the objects differ by the value to which N is initialized before the execution of the. "code" block. Thus, we can now avoid the repetition of the ...

  13. Algorithms

    Indian Academy of Sciences (India)

    algorithms built into the computer corresponding to the logic- circuit rules that are used to .... For the purpose of carrying ou t ari thmetic or logical operations the memory is organized in terms .... In fixed point representation, one essentially uses integer arithmetic operators assuming the binary point to be at some point other ...

  14. An Interactive Personalized Recommendation System Using the Hybrid Algorithm Model

    Directory of Open Access Journals (Sweden)

    Yan Guo

    2017-10-01

    Full Text Available With the rapid development of e-commerce, the contradiction between the disorder of business information and customer demand is increasingly prominent. This study aims to make e-commerce shopping more convenient, and avoid information overload, by an interactive personalized recommendation system using the hybrid algorithm model. The proposed model first uses various recommendation algorithms to get a list of original recommendation results. Combined with the customer’s feedback in an interactive manner, it then establishes the weights of corresponding recommendation algorithms. Finally, the synthetic formula of evidence theory is used to fuse the original results to obtain the final recommendation products. The recommendation performance of the proposed method is compared with that of traditional methods. The results of the experimental study through a Taobao online dress shop clearly show that the proposed method increases the efficiency of data mining in the consumer coverage, the consumer discovery accuracy and the recommendation recall. The hybrid recommendation algorithm complements the advantages of the existing recommendation algorithms in data mining. The interactive assigned-weight method meets consumer demand better and solves the problem of information overload. Meanwhile, our study offers important implications for e-commerce platform providers regarding the design of product recommendation systems.

  15. Efficient Parallel Implementation of Active Appearance Model Fitting Algorithm on GPU

    Directory of Open Access Journals (Sweden)

    Jinwei Wang

    2014-01-01

    Full Text Available The active appearance model (AAM is one of the most powerful model-based object detecting and tracking methods which has been widely used in various situations. However, the high-dimensional texture representation causes very time-consuming computations, which makes the AAM difficult to apply to real-time systems. The emergence of modern graphics processing units (GPUs that feature a many-core, fine-grained parallel architecture provides new and promising solutions to overcome the computational challenge. In this paper, we propose an efficient parallel implementation of the AAM fitting algorithm on GPUs. Our design idea is fine grain parallelism in which we distribute the texture data of the AAM, in pixels, to thousands of parallel GPU threads for processing, which makes the algorithm fit better into the GPU architecture. We implement our algorithm using the compute unified device architecture (CUDA on the Nvidia’s GTX 650 GPU, which has the latest Kepler architecture. To compare the performance of our algorithm with different data sizes, we built sixteen face AAM models of different dimensional textures. The experiment results show that our parallel AAM fitting algorithm can achieve real-time performance for videos even on very high-dimensional textures.

  16. Advanced techniques in reliability model representation and solution

    Science.gov (United States)

    Palumbo, Daniel L.; Nicol, David M.

    1992-01-01

    The current tendency of flight control system designs is towards increased integration of applications and increased distribution of computational elements. The reliability analysis of such systems is difficult because subsystem interactions are increasingly interdependent. Researchers at NASA Langley Research Center have been working for several years to extend the capability of Markov modeling techniques to address these problems. This effort has been focused in the areas of increased model abstraction and increased computational capability. The reliability model generator (RMG) is a software tool that uses as input a graphical object-oriented block diagram of the system. RMG uses a failure-effects algorithm to produce the reliability model from the graphical description. The ASSURE software tool is a parallel processing program that uses the semi-Markov unreliability range evaluator (SURE) solution technique and the abstract semi-Markov specification interface to the SURE tool (ASSIST) modeling language. A failure modes-effects simulation is used by ASSURE. These tools were used to analyze a significant portion of a complex flight control system. The successful combination of the power of graphical representation, automated model generation, and parallel computation leads to the conclusion that distributed fault-tolerant system architectures can now be analyzed.

  17. Application of a single-objective, hybrid genetic algorithm approach to pharmacokinetic model building.

    Science.gov (United States)

    Sherer, Eric A; Sale, Mark E; Pollock, Bruce G; Belani, Chandra P; Egorin, Merrill J; Ivy, Percy S; Lieberman, Jeffrey A; Manuck, Stephen B; Marder, Stephen R; Muldoon, Matthew F; Scher, Howard I; Solit, David B; Bies, Robert R

    2012-08-01

    compounds. The root mean squared error and absolute mean prediction error of the best single-objective hybrid genetic algorithm candidates were a median of 0.2 points higher (range of 38.9 point decrease to 27.3 point increase) and 0.02 points lower (range of 0.98 point decrease to 0.74 point increase), respectively, than that of the final stepwise models. In addition, the best single-objective, hybrid genetic algorithm candidate models had successful convergence and covariance steps for each compound, used the same compartment structure as the manual stepwise approach for 6 of 7 (86 %) compounds, and identified 54 % (7 of 13) of covariates included by the manual stepwise approach and 16 covariate relationships not included by manual stepwise models. The model parameter values between the final manual stepwise and best single-objective, hybrid genetic algorithm models differed by a median of 26.7 % (q₁ = 4.9 % and q₃ = 57.1 %). Finally, the single-objective, hybrid genetic algorithm approach was able to identify models capable of estimating absorption rate parameters for four compounds that the manual stepwise approach did not identify. The single-objective, hybrid genetic algorithm represents a general pharmacokinetic model building methodology whose ability to rapidly search the feasible solution space leads to nearly equivalent or superior model fits to pharmacokinetic data.

  18. Synthetic Optimization Model and Algorithm for Railway Freight Center Station Location and Wagon Flow Organization Problem

    Directory of Open Access Journals (Sweden)

    Xing-cai Liu

    2014-01-01

    Full Text Available The railway freight center stations location and wagon flow organization in railway transport are interconnected, and each of them is complicated in a large-scale rail network. In this paper, a two-stage method is proposed to optimize railway freight center stations location and wagon flow organization together. The location model is present with the objective to minimize the operation cost and fixed construction cost. Then, the second model of wagon flow organization is proposed to decide the optimal train service between different freight center stations. The location of the stations is the output of the first model. A heuristic algorithm that combined tabu search (TS with adaptive clonal selection algorithm (ACSA is proposed to solve those two models. The numerical results show the proposed solution method is effective.

  19. Optimisation of Hidden Markov Model using Baum–Welch algorithm ...

    Indian Academy of Sciences (India)

    s12040-016-0780-0. Optimisation of Hidden Markov Model using. Baum–Welch algorithm for prediction of maximum and minimum temperature over Indian Himalaya. J C Joshi1,∗. , Tankeshwar Kumar2, Sunita Srivastava2 and Divya Sachdeva1.

  20. Optimisation of Hidden Markov Model using Baum–Welch algorithm

    Indian Academy of Sciences (India)

    Home; Journals; Journal of Earth System Science; Volume 126; Issue 1. Optimisation of Hidden Markov Model using Baum–Welch algorithm for prediction of maximum and minimum temperature over Indian Himalaya. J C Joshi Tankeshwar Kumar Sunita Srivastava Divya Sachdeva. Volume 126 Issue 1 February 2017 ...

  1. Epidemic Processes on Complex Networks : Modelling, Simulation and Algorithms

    NARCIS (Netherlands)

    Van de Bovenkamp, R.

    2015-01-01

    Local interactions on a graph will lead to global dynamic behaviour. In this thesis we focus on two types of dynamic processes on graphs: the Susceptible-Infected-Susceptilbe (SIS) virus spreading model, and gossip style epidemic algorithms. The largest part of this thesis is devoted to the SIS

  2. Optimisation of Transfer Function Models using Genetic Algorithms ...

    African Journals Online (AJOL)

    In order to obtain an optimum transfer function estimate, open source software based on genetic algorithm was developed. The software was developed with Visual Basic programming language. In order to test the software, a transfer function model was developed from data obtained from industry. The forecast obtained ...

  3. Stochastic disturbance rejection in model predictive control by randomized algorithms

    NARCIS (Netherlands)

    Batina, Ivo; Stoorvogel, Antonie Arij; Weiland, Siep

    2001-01-01

    In this paper we consider model predictive control with stochastic disturbances and input constraints. We present an algorithm which can solve this problem approximately but with arbitrary high accuracy. The optimization at each time step is a closed loop optimization and therefore takes into

  4. Iteration Capping For Discrete Choice Models Using the EM Algorithm

    NARCIS (Netherlands)

    Kabatek, J.

    2013-01-01

    The Expectation-Maximization (EM) algorithm is a well-established estimation procedure which is used in many domains of econometric analysis. Recent application in a discrete choice framework (Train, 2008) facilitated estimation of latent class models allowing for very exible treatment of unobserved

  5. Evolving the Topology of Hidden Markov Models using Evolutionary Algorithms

    DEFF Research Database (Denmark)

    Thomsen, Réne

    2002-01-01

    Hidden Markov models (HMM) are widely used for speech recognition and have recently gained a lot of attention in the bioinformatics community, because of their ability to capture the information buried in biological sequences. Usually, heuristic algorithms such as Baum-Welch are used to estimate ...

  6. Models and algorithms for Integration of Vehicle and Crew Scheduling

    NARCIS (Netherlands)

    R. Freling (Richard); D. Huisman (Dennis); A.P.M. Wagelmans (Albert)

    2000-01-01

    textabstractThis paper deals with models, relaxations and algorithms for an integrated approach to vehicle and crew scheduling. We discuss potential benefits of integration and provide an overview of the literature, which considers mainly partial integration. Our approach is new in the sense that we

  7. Heterogenous Agents Model with the Worst Out Algorithm

    Czech Academy of Sciences Publication Activity Database

    Vácha, Lukáš; Vošvrda, Miloslav

    -, č. 8 (2006), s. 3-19 ISSN 1801-5999 Institutional research plan: CEZ:AV0Z10750506 Keywords : efficient market hypothesis * fractal market hypothesis * agents' investment horizons * agents' trading strategies * technical trading rules * heterogeneous agent model with stochastic memory * Worst out algorithm Subject RIV: AH - Economics

  8. A sonification algorithm for developing the off-roads models for driving simulators

    Science.gov (United States)

    Chiroiu, Veturia; Brişan, Cornel; Dumitriu, Dan; Munteanu, Ligia

    2018-01-01

    In this paper, a sonification algorithm for developing the off-road models for driving simulators, is proposed. The aim of this algorithm is to overcome difficulties of heuristics identification which are best suited to a particular off-road profile built by measurements. The sonification algorithm is based on the stochastic polynomial chaos analysis suitable in solving equations with random input data. The fluctuations are generated by incomplete measurements leading to inhomogeneities of the cross-sectional curves of off-roads before and after deformation, the unstable contact between the tire and the road and the unreal distribution of contact and friction forces in the unknown contact domains. The approach is exercised on two particular problems and results compare favorably to existing analytical and numerical solutions. The sonification technique represents a useful multiscale analysis able to build a low-cost virtual reality environment with increased degrees of realism for driving simulators and higher user flexibility.

  9. Real Time Optima Tracking Using Harvesting Models of the Genetic Algorithm

    Science.gov (United States)

    Baskaran, Subbiah; Noever, D.

    1999-01-01

    Tracking optima in real time propulsion control, particularly for non-stationary optimization problems is a challenging task. Several approaches have been put forward for such a study including the numerical method called the genetic algorithm. In brief, this approach is built upon Darwinian-style competition between numerical alternatives displayed in the form of binary strings, or by analogy to 'pseudogenes'. Breeding of improved solution is an often cited parallel to natural selection in.evolutionary or soft computing. In this report we present our results of applying a novel model of a genetic algorithm for tracking optima in propulsion engineering and in real time control. We specialize the algorithm to mission profiling and planning optimizations, both to select reduced propulsion needs through trajectory planning and to explore time or fuel conservation strategies.

  10. Fuzzy model predictive control algorithm applied in nuclear power plant

    International Nuclear Information System (INIS)

    Zuheir, Ahmad

    2006-01-01

    The aim of this paper is to design a predictive controller based on a fuzzy model. The Takagi-Sugeno fuzzy model with an Adaptive B-splines neuro-fuzzy implementation is used and incorporated as a predictor in a predictive controller. An optimization approach with a simplified gradient technique is used to calculate predictions of the future control actions. In this approach, adaptation of the fuzzy model using dynamic process information is carried out to build the predictive controller. The easy description of the fuzzy model and the easy computation of the gradient sector during the optimization procedure are the main advantages of the computation algorithm. The algorithm is applied to the control of a U-tube steam generation unit (UTSG) used for electricity generation. (author)

  11. Thickness determination in textile material design: dynamic modeling and numerical algorithms

    International Nuclear Information System (INIS)

    Xu, Dinghua; Ge, Meibao

    2012-01-01

    Textile material design is of paramount importance in the study of functional clothing design. It is therefore important to determine the dynamic heat and moisture transfer characteristics in the human body–clothing–environment system, which directly determine the heat–moisture comfort level of the human body. Based on a model of dynamic heat and moisture transfer with condensation in porous fabric at low temperature, this paper presents a new inverse problem of textile thickness determination (IPTTD). Adopting the idea of the least-squares method, we formulate the IPTTD into a function minimization problem. By means of the finite-difference method, quasi-solution method and direct search method for one-dimensional minimization problems, we construct iterative algorithms of the approximated solution for the IPTTD. Numerical simulation results validate the formulation of the IPTTD and demonstrate the effectiveness of the proposed numerical algorithms. (paper)

  12. Model-based Bayesian signal extraction algorithm for peripheral nerves

    Science.gov (United States)

    Eggers, Thomas E.; Dweiri, Yazan M.; McCallum, Grant A.; Durand, Dominique M.

    2017-10-01

    Objective. Multi-channel cuff electrodes have recently been investigated for extracting fascicular-level motor commands from mixed neural recordings. Such signals could provide volitional, intuitive control over a robotic prosthesis for amputee patients. Recent work has demonstrated success in extracting these signals in acute and chronic preparations using spatial filtering techniques. These extracted signals, however, had low signal-to-noise ratios and thus limited their utility to binary classification. In this work a new algorithm is proposed which combines previous source localization approaches to create a model based method which operates in real time. Approach. To validate this algorithm, a saline benchtop setup was created to allow the precise placement of artificial sources within a cuff and interference sources outside the cuff. The artificial source was taken from five seconds of chronic neural activity to replicate realistic recordings. The proposed algorithm, hybrid Bayesian signal extraction (HBSE), is then compared to previous algorithms, beamforming and a Bayesian spatial filtering method, on this test data. An example chronic neural recording is also analyzed with all three algorithms. Main results. The proposed algorithm improved the signal to noise and signal to interference ratio of extracted test signals two to three fold, as well as increased the correlation coefficient between the original and recovered signals by 10–20%. These improvements translated to the chronic recording example and increased the calculated bit rate between the recovered signals and the recorded motor activity. Significance. HBSE significantly outperforms previous algorithms in extracting realistic neural signals, even in the presence of external noise sources. These results demonstrate the feasibility of extracting dynamic motor signals from a multi-fascicled intact nerve trunk, which in turn could extract motor command signals from an amputee for the end goal of

  13. A quadratic approximation-based algorithm for the solution of multiparametric mixed-integer nonlinear programming problems

    KAUST Repository

    Domínguez, Luis F.

    2012-06-25

    An algorithm for the solution of convex multiparametric mixed-integer nonlinear programming problems arising in process engineering problems under uncertainty is introduced. The proposed algorithm iterates between a multiparametric nonlinear programming subproblem and a mixed-integer nonlinear programming subproblem to provide a series of parametric upper and lower bounds. The primal subproblem is formulated by fixing the integer variables and solved through a series of multiparametric quadratic programming (mp-QP) problems based on quadratic approximations of the objective function, while the deterministic master subproblem is formulated so as to provide feasible integer solutions for the next primal subproblem. To reduce the computational effort when infeasibilities are encountered at the vertices of the critical regions (CRs) generated by the primal subproblem, a simplicial approximation approach is used to obtain CRs that are feasible at each of their vertices. The algorithm terminates when there does not exist an integer solution that is better than the one previously used by the primal problem. Through a series of examples, the proposed algorithm is compared with a multiparametric mixed-integer outer approximation (mp-MIOA) algorithm to demonstrate its computational advantages. © 2012 American Institute of Chemical Engineers (AIChE).

  14. Watermarking Techniques Using Least Significant Bit Algorithm for Digital Image Security Standard Solution- Based Android

    Directory of Open Access Journals (Sweden)

    Ari Muzakir

    2017-05-01

    Full Text Available Ease of deployment of digital image through the internet has positive and negative sides, especially for owners of the original digital image. The positive side of the ease of rapid deployment is the owner of that image deploys digital image files to various sites in the world address. While the downside is that if there is no copyright that serves as protector of the image it will be very easily recognized ownership by other parties. Watermarking is one solution to protect the copyright and know the results of the digital image. With Digital Image Watermarking, copyright resulting digital image will be protected through the insertion of additional information such as owner information and the authenticity of the digital image. The least significant bit (LSB is one of the algorithm is simple and easy to understand. The results of the simulations carried out using android smartphone shows that the LSB watermarking technique is not able to be seen by naked human eye, meaning there is no significant difference in the image of the original files with images that have been inserted watermarking. The resulting image has dimensions of 640x480 with a bit depth of 32 bits. In addition, to determine the function of the ability of the device (smartphone in processing the image using this application used black box testing. 

  15. Performance modeling of parallel algorithms for solving neutron diffusion problems

    International Nuclear Information System (INIS)

    Azmy, Y.Y.; Kirk, B.L.

    1995-01-01

    Neutron diffusion calculations are the most common computational methods used in the design, analysis, and operation of nuclear reactors and related activities. Here, mathematical performance models are developed for the parallel algorithm used to solve the neutron diffusion equation on message passing and shared memory multiprocessors represented by the Intel iPSC/860 and the Sequent Balance 8000, respectively. The performance models are validated through several test problems, and these models are used to estimate the performance of each of the two considered architectures in situations typical of practical applications, such as fine meshes and a large number of participating processors. While message passing computers are capable of producing speedup, the parallel efficiency deteriorates rapidly as the number of processors increases. Furthermore, the speedup fails to improve appreciably for massively parallel computers so that only small- to medium-sized message passing multiprocessors offer a reasonable platform for this algorithm. In contrast, the performance model for the shared memory architecture predicts very high efficiency over a wide range of number of processors reasonable for this architecture. Furthermore, the model efficiency of the Sequent remains superior to that of the hypercube if its model parameters are adjusted to make its processors as fast as those of the iPSC/860. It is concluded that shared memory computers are better suited for this parallel algorithm than message passing computers

  16. ELVIS: Multi-Electrolyte Aqueous Activity Model for Geothermal Solutions

    Science.gov (United States)

    Hingerl, F. F.; Wagner, T.; Driesner, T.; Kulik, D. A.; Kosakowski, G.

    2011-12-01

    High temperature, pressure, and fluid salinities render geochemical modeling of fluid-rock interactions in Enhanced Geothermal Systems a demanding task. Accurate prediction of fluid-mineral equilibria strongly depends on the availability of thermodynamic data and activity models. Typically, the Pitzer activity model is applied for geothermal fluids. A drawback of this model is the large number of parameters required to account for temperature and pressure dependencies, which significantly reduces computational efficiency of reactive transport simulations. In addition, most available parameterizations are valid only at vapor-saturated conditions. As an alternative we implemented the EUNIQUAC local composition model [2] that needs substantially fewer fitting parameters. However, the current EUNIQUAC model design does not include provision for high temperature (>150°C) applications and lacks a formulation for pressure dependence. Therefore, its application to geothermal conditions requires a re-formulation and re-fitting of the model. We developed a new tool termed GEMSFIT that allows generic fitting of activity models (for aqueous electrolyte and non-electrolyte solutions) and equations of state implemented in our geochemical equilibrium solver GEM-Selektor (http://gems.web.psi.ch). GEMSFIT combines a PostgreSQL database for storing and managing the datasets of experimental measurements and interaction parameters, the parallelized genetic algorithm toolbox of MATLAB° for the parameter fitting, and an interface to the numerical kernel of GEM-Selektor to access activity models and perform chemical equilibrium calculations. Benchmarking of the partly re-parameterized EUNIQUAC model against Pitzer revealed that the former is less accurate, which can result in incorrect predictions of mineral precipitation/dissolution. Consequently, we modified the EUNIQUAC model and concurrently introduced a pressure dependence to be able to fit experimental data over wide ranges of

  17. A family of solution algorithms for nonlinear structural analysis based on relaxation equations

    Science.gov (United States)

    Park, K. C.

    1981-01-01

    A family of hierarchical algorithms for nonlinear structural equations are presented. The algorithms are based on the Davidenko-Branin type homotopy and shown to yield consistent hierarchical perturbation equations. The algorithms appear to be particularly suitable to problems involving bifurcation and limit point calculations. An important by-product of the algorithms is that it provides a systematic and economical means for computing the stepsize at each iteration stage when a Newton-like method is employed to solve the systems of equations. Some sample problems are provided to illustrate the characteristics of the algorithms.

  18. Solutions of large-scale electromagnetics problems involving dielectric objects with the parallel multilevel fast multipole algorithm.

    Science.gov (United States)

    Ergül, Özgür

    2011-11-01

    Fast and accurate solutions of large-scale electromagnetics problems involving homogeneous dielectric objects are considered. Problems are formulated with the electric and magnetic current combined-field integral equation and discretized with the Rao-Wilton-Glisson functions. Solutions are performed iteratively by using the multilevel fast multipole algorithm (MLFMA). For the solution of large-scale problems discretized with millions of unknowns, MLFMA is parallelized on distributed-memory architectures using a rigorous technique, namely, the hierarchical partitioning strategy. Efficiency and accuracy of the developed implementation are demonstrated on very large problems involving as many as 100 million unknowns.

  19. The production-distribution problem with order acceptance and package delivery: models and algorithm

    Directory of Open Access Journals (Sweden)

    Khalili Majid

    2016-01-01

    Full Text Available The production planning and distribution are among the most important decisions in the supply chain. Classically, in this problem, it is assumed that all orders have to produced and separately delivered; while, in practice, an order may be rejected if the cost that it brings to the supply chain exceeds its revenue. Moreover, orders can be delivered in a batch to reduce the related costs. This paper considers the production planning and distribution problem with order acceptance and package delivery to maximize the profit. At first, a new mathematical model based on mixed integer linear programming is developed. Using commercial optimization software, the model can optimally solve small or even medium sized instances. For large instances, a solution method, based on imperialist competitive algorithms, is also proposed. Using numerical experiments, the proposed model and algorithm are evaluated.

  20. Modelling and genetic algorithm based optimisation of inverse supply chain

    Science.gov (United States)

    Bányai, T.

    2009-04-01

    (Recycling of household appliances with emphasis on reuse options). The purpose of this paper is the presentation of a possible method for avoiding the unnecessary environmental risk and landscape use through unprovoked large supply chain of collection systems of recycling processes. In the first part of the paper the author presents the mathematical model of recycling related collection systems (applied especially for wastes of electric and electronic products) and in the second part of the work a genetic algorithm based optimisation method will be demonstrated, by the aid of which it is possible to determine the optimal structure of the inverse supply chain from the point of view economical, ecological and logistic objective functions. The model of the inverse supply chain is based on a multi-level, hierarchical collection system. In case of this static model it is assumed that technical conditions are permanent. The total costs consist of three parts: total infrastructure costs, total material handling costs and environmental risk costs. The infrastructure-related costs are dependent only on the specific fixed costs and the specific unit costs of the operation points (collection, pre-treatment, treatment, recycling and reuse plants). The costs of warehousing and transportation are represented by the material handling related costs. The most important factors determining the level of environmental risk cost are the number of out of time recycled (treated or reused) products, the number of supply chain objects and the length of transportation routes. The objective function is the minimization of the total cost taking into consideration the constraints. However a lot of research work discussed the design of supply chain [8], but most of them concentrate on linear cost functions. In the case of this model non-linear cost functions were used. The non-linear cost functions and the possible high number of objects of the inverse supply chain leaded to the problem of choosing a

  1. Solute transport modelling with the variable temporally dependent ...

    Indian Academy of Sciences (India)

    Pintu Das

    2018-02-07

    Feb 7, 2018 ... In this present study, analytical and numerical solutions are obtained for solute transport modelling in homogeneous ..... Clay (0.40). Analytical solution. Numerical solution. Figure 3. Comparison of concentration distribution for sinu- soidal velocity pattern for boundary condition c0. 2 1 ю sec wt р. Ю.

  2. Traveling Wave Solutions in a Reaction-Diffusion Epidemic Model

    OpenAIRE

    Wang, Sheng; Liu, Wenbin; Guo, Zhengguang; Wang, Weiming

    2013-01-01

    We investigate the traveling wave solutions in a reaction-diffusion epidemic model. The existence of the wave solutions is derived through monotone iteration of a pair of classical upper and lower solutions. The traveling wave solutions are shown to be unique and strictly monotonic. Furthermore, we determine the critical minimal wave speed.

  3. Development of solution-gated graphene transistor model for biosensors

    Science.gov (United States)

    Karimi, Hediyeh; Yusof, Rubiyah; Rahmani, Rasoul; Hosseinpour, Hoda; Ahmadi, Mohammad T.

    2014-02-01

    The distinctive properties of graphene, characterized by its high carrier mobility and biocompatibility, have stimulated extreme scientific interest as a promising nanomaterial for future nanoelectronic applications. In particular, graphene-based transistors have been developed rapidly and are considered as an option for DNA sensing applications. Recent findings in the field of DNA biosensors have led to a renewed interest in the identification of genetic risk factors associated with complex human diseases for diagnosis of cancers or hereditary diseases. In this paper, an analytical model of graphene-based solution gated field effect transistors (SGFET) is proposed to constitute an important step towards development of DNA biosensors with high sensitivity and selectivity. Inspired by this fact, a novel strategy for a DNA sensor model with capability of single-nucleotide polymorphism detection is proposed and extensively explained. First of all, graphene-based DNA sensor model is optimized using particle swarm optimization algorithm. Based on the sensing mechanism of DNA sensors, detective parameters ( I ds and V gmin) are suggested to facilitate the decision making process. Finally, the behaviour of graphene-based SGFET is predicted in the presence of single-nucleotide polymorphism with an accuracy of more than 98% which guarantees the reliability of the optimized model for any application of the graphene-based DNA sensor. It is expected to achieve the rapid, quick and economical detection of DNA hybridization which could speed up the realization of the next generation of the homecare sensor system.

  4. Evaluation of different initial solution algorithms to be used in the heuristics optimization to solve the energy resource scheduling in smart grids

    DEFF Research Database (Denmark)

    Sousa, Tiago; Morais, Hugo; Castro, Rui

    2016-01-01

    of finding a final solution near to the optimal than using a random initial solution. This paper proposes two initial solution algorithms to be used by a metaheuristic technique (simulated annealing). These algorithms are tested and evaluated with other published algorithms that obtain initial solution....... The proposed algorithms have been developed as modules to be more flexible their use by other metaheuristics than just simulated annealing. The simulated annealing with different initial solution algorithms has been tested in a 37-bus distribution network with distributed resources, especially electric...... vehicles. The proposed algorithms proved to present results very close to the optimal with a small difference between 0.1%. A deterministic technique is used as comparison and it took around 26 h to obtain the optimal one. On the other hand, the simulated annealing was able of obtaining results around 1...

  5. Models and algorithm of optimization launch and deployment of virtual network functions in the virtual data center

    Science.gov (United States)

    Bolodurina, I. P.; Parfenov, D. I.

    2017-10-01

    The goal of our investigation is optimization of network work in virtual data center. The advantage of modern infrastructure virtualization lies in the possibility to use software-defined networks. However, the existing optimization of algorithmic solutions does not take into account specific features working with multiple classes of virtual network functions. The current paper describes models characterizing the basic structures of object of virtual data center. They including: a level distribution model of software-defined infrastructure virtual data center, a generalized model of a virtual network function, a neural network model of the identification of virtual network functions. We also developed an efficient algorithm for the optimization technology of containerization of virtual network functions in virtual data center. We propose an efficient algorithm for placing virtual network functions. In our investigation we also generalize the well renowned heuristic and deterministic algorithms of Karmakar-Karp.

  6. Fast sweeping algorithm for accurate solution of the TTI eikonal equation using factorization

    KAUST Repository

    bin Waheed, Umair

    2017-06-10

    Traveltime computation is essential for many seismic data processing applications and velocity analysis tools. High-resolution seismic imaging requires eikonal solvers to account for anisotropy whenever it significantly affects the seismic wave kinematics. Moreover, computation of auxiliary quantities, such as amplitude and take-off angle, rely on highly accurate traveltime solutions. However, the finite-difference based eikonal solution for a point-source initial condition has an upwind source-singularity at the source position, since the wavefront curvature is large near the source point. Therefore, all finite-difference solvers, even the high-order ones, show inaccuracies since the errors due to source-singularity spread from the source point to the whole computational domain. We address the source-singularity problem for tilted transversely isotropic (TTI) eikonal solvers using factorization. We solve a sequence of factored tilted elliptically anisotropic (TEA) eikonal equations iteratively, each time by updating the right hand side function. At each iteration, we factor the unknown TEA traveltime into two factors. One of the factors is specified analytically, such that the other factor is smooth in the source neighborhood. Therefore, through the iterative procedure we obtain accurate solution to the TTI eikonal equation. Numerical tests show significant improvement in accuracy due to factorization. The idea can be easily extended to compute accurate traveltimes for models with lower anisotropic symmetries, such as orthorhombic, monoclinic or even triclinic media.

  7. Potts-model grain growth simulations: Parallel algorithms and applications

    Energy Technology Data Exchange (ETDEWEB)

    Wright, S.A.; Plimpton, S.J.; Swiler, T.P. [and others

    1997-08-01

    Microstructural morphology and grain boundary properties often control the service properties of engineered materials. This report uses the Potts-model to simulate the development of microstructures in realistic materials. Three areas of microstructural morphology simulations were studied. They include the development of massively parallel algorithms for Potts-model grain grow simulations, modeling of mass transport via diffusion in these simulated microstructures, and the development of a gradient-dependent Hamiltonian to simulate columnar grain growth. Potts grain growth models for massively parallel supercomputers were developed for the conventional Potts-model in both two and three dimensions. Simulations using these parallel codes showed self similar grain growth and no finite size effects for previously unapproachable large scale problems. In addition, new enhancements to the conventional Metropolis algorithm used in the Potts-model were developed to accelerate the calculations. These techniques enable both the sequential and parallel algorithms to run faster and use essentially an infinite number of grain orientation values to avoid non-physical grain coalescence events. Mass transport phenomena in polycrystalline materials were studied in two dimensions using numerical diffusion techniques on microstructures generated using the Potts-model. The results of the mass transport modeling showed excellent quantitative agreement with one dimensional diffusion problems, however the results also suggest that transient multi-dimension diffusion effects cannot be parameterized as the product of the grain boundary diffusion coefficient and the grain boundary width. Instead, both properties are required. Gradient-dependent grain growth mechanisms were included in the Potts-model by adding an extra term to the Hamiltonian. Under normal grain growth, the primary driving term is the curvature of the grain boundary, which is included in the standard Potts-model Hamiltonian.

  8. Development of algorithm for depreciation costs allocation in dynamic input-output industrial enterprise model

    Directory of Open Access Journals (Sweden)

    Keller Alevtina

    2017-01-01

    Full Text Available The article considers the issue of allocation of depreciation costs in the dynamic inputoutput model of an industrial enterprise. Accounting the depreciation costs in such a model improves the policy of fixed assets management. It is particularly relevant to develop the algorithm for the allocation of depreciation costs in the construction of dynamic input-output model of an industrial enterprise, since such enterprises have a significant amount of fixed assets. Implementation of terms of the adequacy of such an algorithm itself allows: evaluating the appropriateness of investments in fixed assets, studying the final financial results of an industrial enterprise, depending on management decisions in the depreciation policy. It is necessary to note that the model in question for the enterprise is always degenerate. It is caused by the presence of zero rows in the matrix of capital expenditures by lines of structural elements unable to generate fixed assets (part of the service units, households, corporate consumers. The paper presents the algorithm for the allocation of depreciation costs for the model. This algorithm was developed by the authors and served as the basis for further development of the flowchart for subsequent implementation with use of software. The construction of such algorithm and its use for dynamic input-output models of industrial enterprises is actualized by international acceptance of the effectiveness of the use of input-output models for national and regional economic systems. This is what allows us to consider that the solutions discussed in the article are of interest to economists of various industrial enterprises.

  9. Statistical behaviour of adaptive multilevel splitting algorithms in simple models

    International Nuclear Information System (INIS)

    Rolland, Joran; Simonnet, Eric

    2015-01-01

    Adaptive multilevel splitting algorithms have been introduced rather recently for estimating tail distributions in a fast and efficient way. In particular, they can be used for computing the so-called reactive trajectories corresponding to direct transitions from one metastable state to another. The algorithm is based on successive selection–mutation steps performed on the system in a controlled way. It has two intrinsic parameters, the number of particles/trajectories and the reaction coordinate used for discriminating good or bad trajectories. We investigate first the convergence in law of the algorithm as a function of the timestep for several simple stochastic models. Second, we consider the average duration of reactive trajectories for which no theoretical predictions exist. The most important aspect of this work concerns some systems with two degrees of freedom. They are studied in detail as a function of the reaction coordinate in the asymptotic regime where the number of trajectories goes to infinity. We show that during phase transitions, the statistics of the algorithm deviate significatively from known theoretical results when using non-optimal reaction coordinates. In this case, the variance of the algorithm is peaking at the transition and the convergence of the algorithm can be much slower than the usual expected central limit behaviour. The duration of trajectories is affected as well. Moreover, reactive trajectories do not correspond to the most probable ones. Such behaviour disappears when using the optimal reaction coordinate called committor as predicted by the theory. We finally investigate a three-state Markov chain which reproduces this phenomenon and show logarithmic convergence of the trajectory durations

  10. Improving permafrost distribution modelling using feature selection algorithms

    Science.gov (United States)

    Deluigi, Nicola; Lambiel, Christophe; Kanevski, Mikhail

    2016-04-01

    The availability of an increasing number of spatial data on the occurrence of mountain permafrost allows the employment of machine learning (ML) classification algorithms for modelling the distribution of the phenomenon. One of the major problems when dealing with high-dimensional dataset is the number of input features (variables) involved. Application of ML classification algorithms to this large number of variables leads to the risk of overfitting, with the consequence of a poor generalization/prediction. For this reason, applying feature selection (FS) techniques helps simplifying the amount of factors required and improves the knowledge on adopted features and their relation with the studied phenomenon. Moreover, taking away irrelevant or redundant variables from the dataset effectively improves the quality of the ML prediction. This research deals with a comparative analysis of permafrost distribution models supported by FS variable importance assessment. The input dataset (dimension = 20-25, 10 m spatial resolution) was constructed using landcover maps, climate data and DEM derived variables (altitude, aspect, slope, terrain curvature, solar radiation, etc.). It was completed with permafrost evidences (geophysical and thermal data and rock glacier inventories) that serve as training permafrost data. Used FS algorithms informed about variables that appeared less statistically important for permafrost presence/absence. Three different algorithms were compared: Information Gain (IG), Correlation-based Feature Selection (CFS) and Random Forest (RF). IG is a filter technique that evaluates the worth of a predictor by measuring the information gain with respect to the permafrost presence/absence. Conversely, CFS is a wrapper technique that evaluates the worth of a subset of predictors by considering the individual predictive ability of each variable along with the degree of redundancy between them. Finally, RF is a ML algorithm that performs FS as part of its

  11. Development and investigation of an inverse problem solution algorithm for determination of Ap stars magnetic field geometry

    International Nuclear Information System (INIS)

    Piskunov, N.E.

    1985-01-01

    Mathematical formulation of the inverse problem of determination of magnetic field geometry from the polarization profiles of spectral lines is gven. The solving algorithm is proposed. A set of model calculations has shown the effectiveness of the algorithm, the high precision of magnetic star model parameters obtained and also the advantages of the inverse problem method over the commonly used method of interpretation of effective field curves

  12. Optimization of the test intervals of a nuclear safety system by genetic algorithms, solution clustering and fuzzy preference assignment

    Energy Technology Data Exchange (ETDEWEB)

    Zio, E. [Ecole Centrale Paris- Supelec, Paris (France); Bazzo, R. [Politecnico di Milano, Milano (Italy)

    2010-08-15

    In this paper, a procedure is developed for identifying a number of representative solutions manageable for decision-making in a multiobjective optimization problem concerning the test intervals of the components of a safety system of a nuclear power plant. Pareto Front solutions are identified by a genetic algorithm and then clustered by subtractive clustering into 'families'. On the basis of the decision maker's preferences, each family is then synthetically represented by a 'head of the family' solution. This is done by introducing a scoring system that ranks the solutions with respect to the different objectives: a fuzzy preference assignment is employed to this purpose. Level Diagrams are then used to represent, analyze and interpret the Pareto Fronts reduced to the head-of-the-family solutions

  13. Optimization of the test intervals of a nuclear safety system by genetic algorithms, solution clustering and fuzzy preference assignment

    International Nuclear Information System (INIS)

    Zio, E.; Bazzo, R.

    2010-01-01

    In this paper, a procedure is developed for identifying a number of representative solutions manageable for decision-making in a multiobjective optimization problem concerning the test intervals of the components of a safety system of a nuclear power plant. Pareto Front solutions are identified by a genetic algorithm and then clustered by subtractive clustering into 'families'. On the basis of the decision maker's preferences, each family is then synthetically represented by a 'head of the family' solution. This is done by introducing a scoring system that ranks the solutions with respect to the different objectives: a fuzzy preference assignment is employed to this purpose. Level Diagrams are then used to represent, analyze and interpret the Pareto Fronts reduced to the head-of-the-family solutions

  14. A Genetic Algorithm Approach for Modeling a Grounding Electrode

    Science.gov (United States)

    Mishra, Arbind Kumar; Nagaoka, Naoto; Ametani, Akihiro

    This paper has proposed a genetic algorithm based approach to determine a grounding electrode model circuit composed of resistances, inductances and capacitances. The proposed methodology determines the model circuit parameters based on a general ladder circuit directly from a measured result. Transient voltages of some electrodes were measured when applying a step like current. An EMTP simulation of a transient voltage on the grounding electrode has been carried out by adopting the proposed model circuits. The accuracy of the proposed method has been confirmed to be high in comparison with the measured transient voltage.

  15. A comparison of updating algorithms for large $N$ reduced models

    CERN Document Server

    Pérez, Margarita García; Keegan, Liam; Okawa, Masanori; Ramos, Alberto

    2015-01-01

    We investigate Monte Carlo updating algorithms for simulating $SU(N)$ Yang-Mills fields on a single-site lattice, such as for the Twisted Eguchi-Kawai model (TEK). We show that performing only over-relaxation (OR) updates of the gauge links is a valid simulation algorithm for the Fabricius and Haan formulation of this model, and that this decorrelates observables faster than using heat-bath updates. We consider two different methods of implementing the OR update: either updating the whole $SU(N)$ matrix at once, or iterating through $SU(2)$ subgroups of the $SU(N)$ matrix, we find the same critical exponent in both cases, and only a slight difference between the two.

  16. Sustainable logistics and transportation optimization models and algorithms

    CERN Document Server

    Gakis, Konstantinos; Pardalos, Panos

    2017-01-01

    Focused on the logistics and transportation operations within a supply chain, this book brings together the latest models, algorithms, and optimization possibilities. Logistics and transportation problems are examined within a sustainability perspective to offer a comprehensive assessment of environmental, social, ethical, and economic performance measures. Featured models, techniques, and algorithms may be used to construct policies on alternative transportation modes and technologies, green logistics, and incentives by the incorporation of environmental, economic, and social measures. Researchers, professionals, and graduate students in urban regional planning, logistics, transport systems, optimization, supply chain management, business administration, information science, mathematics, and industrial and systems engineering will find the real life and interdisciplinary issues presented in this book informative and useful.

  17. A PISO-like algorithm to simulate superfluid helium flow with the two-fluid model

    CERN Document Server

    Soulaine, Cyprien; Allain, Hervé; Baudouy, Bertrand; Van Weelderen, Rob

    2015-01-01

    This paper presents a segregated algorithm to solve numerically the superfluid helium (He II) equations using the two-fluid model. In order to validate the resulting code and illustrate its potential, different simulations have been performed. First, the flow through a capillary filled with He II with a heated area on one side is simulated and results are compared to analytical solutions in both Landau and Gorter–Mellink flow regimes. Then, transient heat transfer of a forced flow of He II is investigated. Finally, some two-dimensional simulations in a porous medium model are carried out.

  18. Dynamic greedy algorithms for the Edwards-Anderson model

    Science.gov (United States)

    Schnabel, Stefan; Janke, Wolfhard

    2017-11-01

    To provide a novel tool for the investigation of the energy landscape of the Edwards-Anderson spin-glass model we introduce an algorithm that allows an efficient execution of a greedy optimization based on data from a previously performed optimization for a similar configuration. As an application we show how the technique can be used to perform higher-order greedy optimizations and simulated annealing searches with improved performance.

  19. Managing and learning with multiple models: Objectives and optimization algorithms

    Science.gov (United States)

    Probert, William J. M.; Hauser, C.E.; McDonald-Madden, E.; Runge, M.C.; Baxter, P.W.J.; Possingham, H.P.

    2011-01-01

    The quality of environmental decisions should be gauged according to managers' objectives. Management objectives generally seek to maximize quantifiable measures of system benefit, for instance population growth rate. Reaching these goals often requires a certain degree of learning about the system. Learning can occur by using management action in combination with a monitoring system. Furthermore, actions can be chosen strategically to obtain specific kinds of information. Formal decision making tools can choose actions to favor such learning in two ways: implicitly via the optimization algorithm that is used when there is a management objective (for instance, when using adaptive management), or explicitly by quantifying knowledge and using it as the fundamental project objective, an approach new to conservation.This paper outlines three conservation project objectives - a pure management objective, a pure learning objective, and an objective that is a weighted mixture of these two. We use eight optimization algorithms to choose actions that meet project objectives and illustrate them in a simulated conservation project. The algorithms provide a taxonomy of decision making tools in conservation management when there is uncertainty surrounding competing models of system function. The algorithms build upon each other such that their differences are highlighted and practitioners may see where their decision making tools can be improved. ?? 2010 Elsevier Ltd.

  20. Stochastic dynamics modeling solute transport in porous media modeling solute transport in porous media

    CERN Document Server

    Kulasiri, Don

    2002-01-01

    Most of the natural and biological phenomena such as solute transport in porous media exhibit variability which can not be modeled by using deterministic approaches. There is evidence in natural phenomena to suggest that some of the observations can not be explained by using the models which give deterministic solutions. Stochastic processes have a rich repository of objects which can be used to express the randomness inherent in the system and the evolution of the system over time. The attractiveness of the stochastic differential equations (SDE) and stochastic partial differential equations (SPDE) come from the fact that we can integrate the variability of the system along with the scientific knowledge pertaining to the system. One of the aims of this book is to explaim some useufl concepts in stochastic dynamics so that the scientists and engineers with a background in undergraduate differential calculus could appreciate the applicability and appropriateness of these developments in mathematics. The ideas ...

  1. Software Piracy Detection Model Using Ant Colony Optimization Algorithm

    Science.gov (United States)

    Astiqah Omar, Nor; Zakuan, Zeti Zuryani Mohd; Saian, Rizauddin

    2017-06-01

    Internet enables information to be accessible anytime and anywhere. This scenario creates an environment whereby information can be easily copied. Easy access to the internet is one of the factors which contribute towards piracy in Malaysia as well as the rest of the world. According to a survey conducted by Compliance Gap BSA Global Software Survey in 2013 on software piracy, found out that 43 percent of the software installed on PCs around the world was not properly licensed, the commercial value of the unlicensed installations worldwide was reported to be 62.7 billion. Piracy can happen anywhere including universities. Malaysia as well as other countries in the world is faced with issues of piracy committed by the students in universities. Piracy in universities concern about acts of stealing intellectual property. It can be in the form of software piracy, music piracy, movies piracy and piracy of intellectual materials such as books, articles and journals. This scenario affected the owner of intellectual property as their property is in jeopardy. This study has developed a classification model for detecting software piracy. The model was developed using a swarm intelligence algorithm called the Ant Colony Optimization algorithm. The data for training was collected by a study conducted in Universiti Teknologi MARA (Perlis). Experimental results show that the model detection accuracy rate is better as compared to J48 algorithm.

  2. Periodic Solutions for a Delayed Population Model on Time Scales

    OpenAIRE

    Kejun Zhuang; Zhaohui Wen

    2010-01-01

    This paper deals with a delayed single population model on time scales. With the assistance of coincidence degree theory, sufficient conditions for existence of periodic solutions are obtained. Furthermore, the better estimations for bounds of periodic solutions are established.

  3. Modified Hyperspheres Algorithm to Trace Homotopy Curves of Nonlinear Circuits Composed by Piecewise Linear Modelled Devices

    Directory of Open Access Journals (Sweden)

    H. Vazquez-Leal

    2014-01-01

    Full Text Available We present a homotopy continuation method (HCM for finding multiple operating points of nonlinear circuits composed of devices modelled by using piecewise linear (PWL representations. We propose an adaptation of the modified spheres path tracking algorithm to trace the homotopy trajectories of PWL circuits. In order to assess the benefits of this proposal, four nonlinear circuits composed of piecewise linear modelled devices are analysed to determine their multiple operating points. The results show that HCM can find multiple solutions within a single homotopy trajectory. Furthermore, we take advantage of the fact that homotopy trajectories are PWL curves meant to replace the multidimensional interpolation and fine tuning stages of the path tracking algorithm with a simple and highly accurate procedure based on the parametric straight line equation.

  4. A robust model predictive control algorithm for uncertain nonlinear systems that guarantees resolvability

    Science.gov (United States)

    Acikmese, Ahmet Behcet; Carson, John M., III

    2006-01-01

    A robustly stabilizing MPC (model predictive control) algorithm for uncertain nonlinear systems is developed that guarantees resolvability. With resolvability, initial feasibility of the finite-horizon optimal control problem implies future feasibility in a receding-horizon framework. The control consists of two components; (i) feed-forward, and (ii) feedback part. Feed-forward control is obtained by online solution of a finite-horizon optimal control problem for the nominal system dynamics. The feedback control policy is designed off-line based on a bound on the uncertainty in the system model. The entire controller is shown to be robustly stabilizing with a region of attraction composed of initial states for which the finite-horizon optimal control problem is feasible. The controller design for this algorithm is demonstrated on a class of systems with uncertain nonlinear terms that have norm-bounded derivatives and derivatives in polytopes. An illustrative numerical example is also provided.

  5. Small Body GN&C Research Report: A Robust Model Predictive Control Algorithm with Guaranteed Resolvability

    Science.gov (United States)

    Acikmese, Behcet A.; Carson, John M., III

    2005-01-01

    A robustly stabilizing MPC (model predictive control) algorithm for uncertain nonlinear systems is developed that guarantees the resolvability of the associated finite-horizon optimal control problem in a receding-horizon implementation. The control consists of two components; (i) feedforward, and (ii) feedback part. Feed-forward control is obtained by online solution of a finite-horizon optimal control problem for the nominal system dynamics. The feedback control policy is designed off-line based on a bound on the uncertainty in the system model. The entire controller is shown to be robustly stabilizing with a region of attraction composed of initial states for which the finite-horizon optimal control problem is feasible. The controller design for this algorithm is demonstrated on a class of systems with uncertain nonlinear terms that have norm-bounded derivatives, and derivatives in polytopes. An illustrative numerical example is also provided.

  6. Hybrid algorithm: A cost efficient solution for ONU placement in Fiber-Wireless (FiWi) network

    Science.gov (United States)

    Bhatt, Uma Rathore; Chouhan, Nitin; Upadhyay, Raksha

    2015-03-01

    Fiber-Wireless (FiWi) network is a promising access technology as it integrates the technical merits of optical and wireless access networks. FiWi provides large bandwidth and high stability of optical network and lower cost of wireless network respectively. Therefore, FiWi gives users to access broadband services in an "anywhere-anytime" way. One of the key issues in FiWi network is its deployment cost, which depends on the number of ONUs in the network. Therefore optimal placement of ONUs is desirable to design a cost effective network. In this paper, we propose an algorithm for optimal placement of ONUs. First we place an ONU in the center of each grid then we form a set of wireless routers associated with each ONU according to wireless hop number. The number of ONUs are minimized in such a way, that all the wireless routers can communicate to at least one of the ONUs. The number of ONUs in the network further reduced by using genetic algorithm. The effectiveness of the proposed algorithm is tested by considering Internet traffic as well as peer-to-peer (p2p) traffic in the network, which is a current need. Simulation results show that the proposed algorithm is better than existing algorithms in minimizing number of ONUs in the network for both types of traffics. Hence proposed algorithm offers cost effective solution to design the FiWi network.

  7. Exact and approximate Fourier rebinning algorithms for the solution of the data truncation problem in 3-D PET.

    Science.gov (United States)

    Bouallègue, Fayçal Ben; Crouzet, Jean-François; Comtat, Claude; Fourcade, Marjolaine; Mohammadi, Bijan; Mariano-Goulart, Denis

    2007-07-01

    This paper presents an extended 3-D exact rebinning formula in the Fourier space that leads to an iterative reprojection algorithm (iterative FOREPROJ), which enables the estimation of unmeasured oblique projection data on the basis of the whole set of measured data. In first approximation, this analytical formula also leads to an extended Fourier rebinning equation that is the basis for an approximate reprojection algorithm (extended FORE). These algorithms were evaluated on numerically simulated 3-D positron emission tomography (PET) data for the solution of the truncation problem, i.e., the estimation of the missing portions in the oblique projection data, before the application of algorithms that require complete projection data such as some rebinning methods (FOREX) or 3-D reconstruction algorithms (3DRP or direct Fourier methods). By taking advantage of all the 3-D data statistics, the iterative FOREPROJ reprojection provides a reliable alternative to the classical FOREPROJ method, which only exploits the low-statistics nonoblique data. It significantly improves the quality of the external reconstructed slices without loss of spatial resolution. As for the approximate extended FORE algorithm, it clearly exhibits limitations due to axial interpolations, but will require clinical studies with more realistic measured data in order to decide on its pertinence.

  8. Classical and Weak Solutions for Two Models in Mathematical Finance

    Science.gov (United States)

    Gyulov, Tihomir B.; Valkov, Radoslav L.

    2011-12-01

    We study two mathematical models, arising in financial mathematics. These models are one-dimensional analogues of the famous Black-Scholes equation on finite interval. The main difficulty is the degeneration at the both ends of the space interval. First, classical solutions are studied. Positivity and convexity properties of the solutions are discussed. Variational formulation in weighted Sobolev spaces is introduced and existence and uniqueness of the weak solution is proved. Maximum principle for weak solution is discussed.

  9. Motion Model Employment using interacting Motion Model Algorithm

    DEFF Research Database (Denmark)

    Hussain, Dil Muhammad Akbar

    2006-01-01

    model being correct is computed through a likelihood function for each model.  The study presented a simple technique to introduce additional models into the system using deterministic acceleration which basically defines the dynamics of the system.  Therefore, based on this value more motion models can...

  10. Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments

    Science.gov (United States)

    Jiang, Yanhua; Xiong, Guangming; Chen, Huiyan; Lee, Dah-Jye

    2014-01-01

    This paper presents a monocular visual odometry algorithm that incorporates a wheeled vehicle model for ground vehicles. The main innovation of this algorithm is to use the single-track bicycle model to interpret the relationship between the yaw rate and side slip angle, which are the two most important parameters that describe the motion of a wheeled vehicle. Additionally, the pitch angle is also considered since the planar-motion hypothesis often fails due to the dynamic characteristics of wheel suspensions and tires in real-world environments. Linearization is used to calculate a closed-form solution of the motion parameters that works as a hypothesis generator in a RAndom SAmple Consensus (RANSAC) scheme to reduce the complexity in solving equations involving trigonometric. All inliers found are used to refine the winner solution through minimizing the reprojection error. Finally, the algorithm is applied to real-time on-board visual localization applications. Its performance is evaluated by comparing against the state-of-the-art monocular visual odometry methods using both synthetic data and publicly available datasets over several kilometers in dynamic outdoor environments. PMID:25256109

  11. Incorporating a wheeled vehicle model in a new monocular visual odometry algorithm for dynamic outdoor environments.

    Science.gov (United States)

    Jiang, Yanhua; Xiong, Guangming; Chen, Huiyan; Lee, Dah-Jye

    2014-09-01

    This paper presents a monocular visual odometry algorithm that incorporates a wheeled vehicle model for ground vehicles. The main innovation of this algorithm is to use the single-track bicycle model to interpret the relationship between the yaw rate and side slip angle, which are the two most important parameters that describe the motion of a wheeled vehicle. Additionally, the pitch angle is also considered since the planar-motion hypothesis often fails due to the dynamic characteristics of wheel suspensions and tires in real-world environments. Linearization is used to calculate a closed-form solution of the motion parameters that works as a hypothesis generator in a RAndom SAmple Consensus (RANSAC) scheme to reduce the complexity in solving equations involving trigonometric. All inliers found are used to refine the winner solution through minimizing the reprojection error. Finally, the algorithm is applied to real-time on-board visual localization applications. Its performance is evaluated by comparing against the state-of-the-art monocular visual odometry methods using both synthetic data and publicly available datasets over several kilometers in dynamic outdoor environments.

  12. Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments

    Directory of Open Access Journals (Sweden)

    Yanhua Jiang

    2014-09-01

    Full Text Available This paper presents a monocular visual odometry algorithm that incorporates a wheeled vehicle model for ground vehicles. The main innovation of this algorithm is to use the single-track bicycle model to interpret the relationship between the yaw rate and side slip angle, which are the two most important parameters that describe the motion of a wheeled vehicle. Additionally, the pitch angle is also considered since the planar-motion hypothesis often fails due to the dynamic characteristics of wheel suspensions and tires in real-world environments. Linearization is used to calculate a closed-form solution of the motion parameters that works as a hypothesis generator in a RAndom SAmple Consensus (RANSAC scheme to reduce the complexity in solving equations involving trigonometric. All inliers found are used to refine the winner solution through minimizing the reprojection error. Finally, the algorithm is applied to real-time on-board visual localization applications. Its performance is evaluated by comparing against the state-of-the-art monocular visual odometry methods using both synthetic data and publicly available datasets over several kilometers in dynamic outdoor environments.

  13. Earthquake forecast models for Italy based on the RI algorithm

    Directory of Open Access Journals (Sweden)

    Kazuyoshi Z. Nanjo

    2010-11-01

    Full Text Available This study provides an overview of relative-intensity (RI-based earthquake forecast models that have been submitted for the 5-year and 10-year testing classes and the 3-month class of the Italian experiment within the Collaboratory for the Study of Earthquake Predictability (CSEP. The RI algorithm starts as a binary forecast system based on the working assumption that future large earthquakes are considered likely to occur at sites of higher seismic activity in the past. The measure of RI is the simply counting of the number of past earthquakes, which is known as the RI of seismicity. To improve the RI forecast performance, we first expand the RI algorithm to become part of a general class of smoothed seismicity models. We then convert the RI representation from a binary system into a testable CSEP model that forecasts the numbers of earthquakes for the predefined magnitudes. Our parameter tuning for the CSEP models is based on the past seismicity. The final submission is a set of two numerical data files that were created by tuned 5-year and 10-year models and an executable computer code of a tuned 3-month model, to examine which testing class is more meaningful in terms of the RI hypothesis. The main purpose of our participation is to better understand the importance (or lack of importance of RI of seismicity for earthquake forecastability.

  14. Optimisation-Based Solution Methods for Set Partitioning Models

    DEFF Research Database (Denmark)

    Rasmussen, Matias Sevel

    The scheduling of crew, i.e. the construction of work schedules for crew members, is often not a trivial task, but a complex puzzle. The task is complicated by rules, restrictions, and preferences. Therefore, manual solutions as well as solutions from standard software packages are not always su......_cient with respect to solution quality and solution time. Enhancement of the overall solution quality as well as the solution time can be of vital importance to many organisations. The _elds of operations research and mathematical optimisation deal with mathematical modelling of di_cult scheduling problems (among...... other topics). The _elds also deal with the development of sophisticated solution methods for these mathematical models. This thesis describes the set partitioning model which has been widely used for modelling crew scheduling problems. Integer properties for the set partitioning model are shown...

  15. Development and evaluation of thermal model reduction algorithms for spacecraft

    Science.gov (United States)

    Deiml, Michael; Suderland, Martin; Reiss, Philipp; Czupalla, Markus

    2015-05-01

    This paper is concerned with the topic of the reduction of thermal models of spacecraft. The work presented here has been conducted in cooperation with the company OHB AG, formerly Kayser-Threde GmbH, and the Institute of Astronautics at Technische Universität München with the goal to shorten and automatize the time-consuming and manual process of thermal model reduction. The reduction of thermal models can be divided into the simplification of the geometry model for calculation of external heat flows and radiative couplings and into the reduction of the underlying mathematical model. For simplification a method has been developed which approximates the reduced geometry model with the help of an optimization algorithm. Different linear and nonlinear model reduction techniques have been evaluated for their applicability in reduction of the mathematical model. Thereby the compatibility with the thermal analysis tool ESATAN-TMS is of major concern, which restricts the useful application of these methods. Additional model reduction methods have been developed, which account to these constraints. The Matrix Reduction method allows the approximation of the differential equation to reference values exactly expect for numerical errors. The summation method enables a useful, applicable reduction of thermal models that can be used in industry. In this work a framework for model reduction of thermal models has been created, which can be used together with a newly developed graphical user interface for the reduction of thermal models in industry.

  16. A Multiple Model Prediction Algorithm for CNC Machine Wear PHM

    Directory of Open Access Journals (Sweden)

    Huimin Chen

    2011-01-01

    Full Text Available The 2010 PHM data challenge focuses on the remaining useful life (RUL estimation for cutters of a high speed CNC milling machine using measurements from dynamometer, accelerometer, and acoustic emission sensors. We present a multiple model approach for wear depth estimation of milling machine cutters using the provided data. The feature selection, initial wear estimation and multiple model fusion components of the proposed algorithm are explained in details and compared with several alternative methods using the training data. The final submission ranked #2 among professional and student participants and the method is applicable to other data driven PHM problems.

  17. Linguistically motivated statistical machine translation models and algorithms

    CERN Document Server

    Xiong, Deyi

    2015-01-01

    This book provides a wide variety of algorithms and models to integrate linguistic knowledge into Statistical Machine Translation (SMT). It helps advance conventional SMT to linguistically motivated SMT by enhancing the following three essential components: translation, reordering and bracketing models. It also serves the purpose of promoting the in-depth study of the impacts of linguistic knowledge on machine translation. Finally it provides a systematic introduction of Bracketing Transduction Grammar (BTG) based SMT, one of the state-of-the-art SMT formalisms, as well as a case study of linguistically motivated SMT on a BTG-based platform.

  18. Comparison of evolutionary algorithms in gene regulatory network model inference.

    LENUS (Irish Health Repository)

    2010-01-01

    ABSTRACT: BACKGROUND: The evolution of high throughput technologies that measure gene expression levels has created a data base for inferring GRNs (a process also known as reverse engineering of GRNs). However, the nature of these data has made this process very difficult. At the moment, several methods of discovering qualitative causal relationships between genes with high accuracy from microarray data exist, but large scale quantitative analysis on real biological datasets cannot be performed, to date, as existing approaches are not suitable for real microarray data which are noisy and insufficient. RESULTS: This paper performs an analysis of several existing evolutionary algorithms for quantitative gene regulatory network modelling. The aim is to present the techniques used and offer a comprehensive comparison of approaches, under a common framework. Algorithms are applied to both synthetic and real gene expression data from DNA microarrays, and ability to reproduce biological behaviour, scalability and robustness to noise are assessed and compared. CONCLUSIONS: Presented is a comparison framework for assessment of evolutionary algorithms, used to infer gene regulatory networks. Promising methods are identified and a platform for development of appropriate model formalisms is established.

  19. A new model and simple algorithms for multi-label mumford-shah problems

    KAUST Repository

    Hong, Byungwoo

    2013-06-01

    In this work, we address the multi-label Mumford-Shah problem, i.e., the problem of jointly estimating a partitioning of the domain of the image, and functions defined within regions of the partition. We create algorithms that are efficient, robust to undesirable local minima, and are easy-to-implement. Our algorithms are formulated by slightly modifying the underlying statistical model from which the multi-label Mumford-Shah functional is derived. The advantage of this statistical model is that the underlying variables: the labels and the functions are less coupled than in the original formulation, and the labels can be computed from the functions with more global updates. The resulting algorithms can be tuned to the desired level of locality of the solution: from fully global updates to more local updates. We demonstrate our algorithm on two applications: joint multi-label segmentation and denoising, and joint multi-label motion segmentation and flow estimation. We compare to the state-of-the-art in multi-label Mumford-Shah problems and show that we achieve more promising results. © 2013 IEEE.

  20. A parallel domain decomposition algorithm for coastal ocean circulation models based on integer linear programming

    Science.gov (United States)

    Jordi, Antoni; Georgas, Nickitas; Blumberg, Alan

    2017-05-01

    This paper presents a new parallel domain decomposition algorithm based on integer linear programming (ILP), a mathematical optimization method. To minimize the computation time of coastal ocean circulation models, the ILP decomposition algorithm divides the global domain in local domains with balanced work load according to the number of processors and avoids computations over as many as land grid cells as possible. In addition, it maintains the use of logically rectangular local domains and achieves the exact same results as traditional domain decomposition algorithms (such as Cartesian decomposition). However, the ILP decomposition algorithm may not converge to an exact solution for relatively large domains. To overcome this problem, we developed two ILP decomposition formulations. The first one (complete formulation) has no additional restriction, although it is impractical for large global domains. The second one (feasible) imposes local domains with the same dimensions and looks for the feasibility of such decomposition, which allows much larger global domains. Parallel performance of both ILP formulations is compared to a base Cartesian decomposition by simulating two cases with the newly created parallel version of the Stevens Institute of Technology's Estuarine and Coastal Ocean Model (sECOM). Simulations with the ILP formulations run always faster than the ones with the base decomposition, and the complete formulation is better than the feasible one when it is applicable. In addition, parallel efficiency with the ILP decomposition may be greater than one.

  1. Theoretical modelling of actinide spectra in solution

    International Nuclear Information System (INIS)

    Danilo, Cecile

    2009-01-01

    The framework of this PhD is the interpretation of Nuclear Magnetic Relaxation Dispersion experiments performed on solvated U 4+ , NpO 2 + and PuO 2 2+ , which all have a f 2 configuration. Unexpectedly the two actinyl ions have a much higher relaxivity than U 4+ ,. One possible explanation is that the electronic relaxation rate is faster for Uranium(IV) than for the actinyl ions. We address this problem by exploring the electronic spectrum of the three compounds in gas phase and in solution with a two-step SOCI (Spin-Orbit Configuration-Interaction) method. The influence of electron correlation (treated in the first step) and spin-orbit relaxation effects (considered in the second step) has been discussed thoroughly. Solvent effects have been investigated as well. Another issue that has been questioned is the accuracy of Density Functional Theory for the study of actinide species. This matter has been discussed by comparing its performance to wave-function based correlated methods. The chemical problem chosen was the water exchange in [UO 2 2+ (H 2 O) 5 ]. We looked at the associative and at the dissociative mechanisms using a model with one additional water in the second hydration sphere. The last part of the thesis dealt with the spectroscopy of coordinated Uranyl(V). Absorption spectrum of Uranyl(V) with various ligands has been recorded. The first sharp absorption bands in the Near-Infrared region were assigned to the Uranium centered 5f-5f transitions, but uncertainties remained for the assignment of transitions observed in the Visible region. We computed the spectra of naked UO 2 + and [UO 2 (CO 3 ) 3 ] 5- to elucidate the spectral changes induced by the carbonate ligands. (author) [fr

  2. Algorithm of DRM with Kinetic Damping for Finite Element Static Solution of Strain-Softening Structures

    Directory of Open Access Journals (Sweden)

    Wei Wang

    2017-01-01

    Full Text Available In order to deal with the divergence and instability due to the ill-posedness of the nonlinear finite element (FE model of strain-softening structure in implicit static analysis, the dynamic relaxation method (DRM was used with kinetic damping to solve the static increments in the incremental solution procedure so that the problem becomes well-posed. Moreover, in DRM there is no need to assemble and inverse the stiffness matrix as in implicit static analysis such that the associated computational cost is avoided. The ascending branch of static equilibrium path was solved by load increments, while the peak point and the descending branch were solved by displacement increments. Two numerical examples illustrated the effectiveness of such application of DRM in the FE analysis of static equilibrium path of strain-softening structures.

  3. Global identifiability of linear compartmental models--a computer algebra algorithm.

    Science.gov (United States)

    Audoly, S; D'Angiò, L; Saccomani, M P; Cobelli, C

    1998-01-01

    A priori global identifiability deals with the uniqueness of the solution for the unknown parameters of a model and is, thus, a prerequisite for parameter estimation of biological dynamic models. Global identifiability is however difficult to test, since it requires solving a system of algebraic nonlinear equations which increases both in nonlinearity degree and number of terms and unknowns with increasing model order. In this paper, a computer algebra tool, GLOBI (GLOBal Identifiability) is presented, which combines the topological transfer function method with the Buchberger algorithm, to test global identifiability of linear compartmental models. GLOBI allows for the automatic testing of a priori global identifiability of general structure compartmental models from general multi input-multi output experiments. Examples of usage of GLOBI to analyze a priori global identifiability of some complex biological compartmental models are provided.

  4. High speed railway track dynamics models, algorithms and applications

    CERN Document Server

    Lei, Xiaoyan

    2017-01-01

    This book systematically summarizes the latest research findings on high-speed railway track dynamics, made by the author and his research team over the past decade. It explores cutting-edge issues concerning the basic theory of high-speed railways, covering the dynamic theories, models, algorithms and engineering applications of the high-speed train and track coupling system. Presenting original concepts, systematic theories and advanced algorithms, the book places great emphasis on the precision and completeness of its content. The chapters are interrelated yet largely self-contained, allowing readers to either read through the book as a whole or focus on specific topics. It also combines theories with practice to effectively introduce readers to the latest research findings and developments in high-speed railway track dynamics. It offers a valuable resource for researchers, postgraduates and engineers in the fields of civil engineering, transportation, highway & railway engineering.

  5. A model of algorithmic representation of a business process

    Directory of Open Access Journals (Sweden)

    E. I. Koshkarova

    2014-01-01

    Full Text Available This article presents and justifies the possibility of developing a method for estimation and optimization of an enterprise business processes; the proposed method is based on identity of two notions – an algorithm and a business process. The described method relies on extraction of a recursive model from the business process, based on the example of one process automated by the BPM system and further estimation and optimization of that process in accordance with estimation and optimization techniques applied to algorithms. The results of this investigation could be used by experts working in the field of reengineering of enterprise business processes, automation of business processes along with development of enterprise informational systems.

  6. Direct and iterative algorithms for the parallel solution of the one-dimensional macroscopic Navier-Stokes equations

    International Nuclear Information System (INIS)

    Doster, J.M.; Sills, E.D.

    1986-01-01

    Current efforts are under way to develop and evaluate numerical algorithms for the parallel solution of the large sparse matrix equations associated with the finite difference representation of the macroscopic Navier-Stokes equations. Previous work has shown that these equations can be cast into smaller coupled matrix equations suitable for solution utilizing multiple computer processors operating in parallel. The individual processors themselves may exhibit parallelism through the use of vector pipelines. This wor, has concentrated on the one-dimensional drift flux form of the Navier-Stokes equations. Direct and iterative algorithms that may be suitable for implementation on parallel computer architectures are evaluated in terms of accuracy and overall execution speed. This work has application to engineering and training simulations, on-line process control systems, and engineering workstations where increased computational speeds are required

  7. Model order reduction using eigen algorithm | Singh | International ...

    African Journals Online (AJOL)

    -scale dynamic systems where denominator polynomial determined through Eigen algorithm and numerator polynomial via factor division algorithm. In Eigen algorithm, the most dominant Eigen value of both original and reduced order ...

  8. A self-organizing algorithm for modeling protein loops.

    Directory of Open Access Journals (Sweden)

    Pu Liu

    2009-08-01

    Full Text Available Protein loops, the flexible short segments connecting two stable secondary structural units in proteins, play a critical role in protein structure and function. Constructing chemically sensible conformations of protein loops that seamlessly bridge the gap between the anchor points without introducing any steric collisions remains an open challenge. A variety of algorithms have been developed to tackle the loop closure problem, ranging from inverse kinematics to knowledge-based approaches that utilize pre-existing fragments extracted from known protein structures. However, many of these approaches focus on the generation of conformations that mainly satisfy the fixed end point condition, leaving the steric constraints to be resolved in subsequent post-processing steps. In the present work, we describe a simple solution that simultaneously satisfies not only the end point and steric conditions, but also chirality and planarity constraints. Starting from random initial atomic coordinates, each individual conformation is generated independently by using a simple alternating scheme of pairwise distance adjustments of randomly chosen atoms, followed by fast geometric matching of the conformationally rigid components of the constituent amino acids. The method is conceptually simple, numerically stable and computationally efficient. Very importantly, additional constraints, such as those derived from NMR experiments, hydrogen bonds or salt bridges, can be incorporated into the algorithm in a straightforward and inexpensive way, making the method ideal for solving more complex multi-loop problems. The remarkable performance and robustness of the algorithm are demonstrated on a set of protein loops of length 4, 8, and 12 that have been used in previous studies.

  9. Modelling Paleoearthquake Slip Distributions using a Gentic Algorithm

    Science.gov (United States)

    Lindsay, Anthony; Simão, Nuno; McCloskey, John; Nalbant, Suleyman; Murphy, Shane; Bhloscaidh, Mairead Nic

    2013-04-01

    Along the Sunda trench, the annual growth rings of coral microatolls store long term records of tectonic deformation. Spread over large areas of an active megathrust fault, they offer the possibility of high resolution reconstructions of slip for a number of paleo-earthquakes. These data are complex with spatial and temporal variations in uncertainty. Rather than assuming that any one model will uniquely fit the data, Monte Carlo Slip Estimation (MCSE) modelling produces a catalogue of possible models for each event. From each earthquake's catalogue, a model is selected and a possible history of slip along the fault reconstructed. By generating multiple histories, then finding the average slip during each earthquake, a probabilistic history of slip along the fault can be generated and areas that may have a large slip deficit identified. However, the MCSE technique requires the production of many hundreds of billions of models to yield the few models that fit the observed coral data. In an attempt to accelerate this process, we have designed a Genetic Algorithm (GA). The GA uses evolutionary operators to recombine the information held by a population of possible slip models to produce a set of new models, based on how well they reproduce a set of coral deformation data. Repeated iterations of the algorithm produce populations of improved models, each generation better satisfying the coral data. Preliminary results have shown the GA to be capable of recovering synthetically generated slip distributions based their displacements of sets of corals faster than the MCSE technique. The results of the systematic testing of the GA technique and its performance using both synthetic and observed coral displacement data will be presented.

  10. Periodic solutions of nonautonomous differential systems modeling obesity population

    International Nuclear Information System (INIS)

    Arenas, Abraham J.; Gonzalez-Parra, Gilberto; Jodar, Lucas

    2009-01-01

    In this paper we study the periodic behaviour of the solutions of a nonautonomous model for obesity population. The mathematical model represented by a nonautonomous system of nonlinear ordinary differential equations is used to model the dynamics of obese populations. Numerical simulations suggest periodic behaviour of subpopulations solutions. Sufficient conditions which guarantee the existence of a periodic positive solution are obtained using a continuation theorem based on coincidence degree theory.

  11. Periodic solutions of nonautonomous differential systems modeling obesity population

    Energy Technology Data Exchange (ETDEWEB)

    Arenas, Abraham J. [Departamento de Matematicas y Estadistica, Universidad de Cordoba Monteria (Colombia)], E-mail: aarenas@sinu.unicordoba.edu.co; Gonzalez-Parra, Gilberto [Departamento de Calculo, Universidad de los Andes, Merida (Venezuela, Bolivarian Republic of)], E-mail: gcarlos@ula.ve; Jodar, Lucas [Instituto de Matematica Multidisciplinar, Universidad Politecnica de Valencia Edificio 8G, 2o, 46022 Valencia (Spain)], E-mail: ljodar@imm.upv.es

    2009-10-30

    In this paper we study the periodic behaviour of the solutions of a nonautonomous model for obesity population. The mathematical model represented by a nonautonomous system of nonlinear ordinary differential equations is used to model the dynamics of obese populations. Numerical simulations suggest periodic behaviour of subpopulations solutions. Sufficient conditions which guarantee the existence of a periodic positive solution are obtained using a continuation theorem based on coincidence degree theory.

  12. A universal optimization strategy for ant colony optimization algorithms based on the Physarum-inspired mathematical model

    International Nuclear Information System (INIS)

    Zhang, Zili; Gao, Chao; Liu, Yuxin; Qian, Tao

    2014-01-01

    Ant colony optimization (ACO) algorithms often fall into the local optimal solution and have lower search efficiency for solving the travelling salesman problem (TSP). According to these shortcomings, this paper proposes a universal optimization strategy for updating the pheromone matrix in the ACO algorithms. The new optimization strategy takes advantages of the unique feature of critical paths reserved in the process of evolving adaptive networks of the Physarum-inspired mathematical model (PMM). The optimized algorithms, denoted as PMACO algorithms, can enhance the amount of pheromone in the critical paths and promote the exploitation of the optimal solution. Experimental results in synthetic and real networks show that the PMACO algorithms are more efficient and robust than the traditional ACO algorithms, which are adaptable to solve the TSP with single or multiple objectives. Meanwhile, we further analyse the influence of parameters on the performance of the PMACO algorithms. Based on these analyses, the best values of these parameters are worked out for the TSP. (paper)

  13. Modelling of Hydrothermal Unit Commitment Coordination Using Efficient Metaheuristic Algorithm: A Hybridized Approach

    Directory of Open Access Journals (Sweden)

    Suman Sutradhar

    2016-01-01

    Full Text Available In this paper, a novel approach of hybridization of two efficient metaheuristic algorithms is proposed for energy system analysis and modelling based on a hydro and thermal based power system in both single and multiobjective environment. The scheduling of hydro and thermal power is modelled descriptively including the handling method of various practical nonlinear constraints. The main goal for the proposed modelling is to minimize the total production cost (which is highly nonlinear and nonconvex problem and emission while satisfying involved hydro and thermal unit commitment limitations. The cascaded hydro reservoirs of hydro subsystem and intertemporal constraints regarding thermal units along with nonlinear nonconvex, mixed-integer mixed-binary objective function make the search space highly complex. To solve such a complicated system, a hybridization of Gray Wolf Optimization and Artificial Bee Colony algorithm, that is, h-ABC/GWO, is used for better exploration and exploitation in the multidimensional search space. Two different test systems are used for modelling and analysis. Experimental results demonstrate the superior performance of the proposed algorithm as compared to other recently reported ones in terms of convergence and better quality of solutions.

  14. Exploration Of Deep Learning Algorithms Using Openacc Parallel Programming Model

    KAUST Repository

    Hamam, Alwaleed A.

    2017-03-13

    Deep learning is based on a set of algorithms that attempt to model high level abstractions in data. Specifically, RBM is a deep learning algorithm that used in the project to increase it\\'s time performance using some efficient parallel implementation by OpenACC tool with best possible optimizations on RBM to harness the massively parallel power of NVIDIA GPUs. GPUs development in the last few years has contributed to growing the concept of deep learning. OpenACC is a directive based ap-proach for computing where directives provide compiler hints to accelerate code. The traditional Restricted Boltzmann Ma-chine is a stochastic neural network that essentially perform a binary version of factor analysis. RBM is a useful neural net-work basis for larger modern deep learning model, such as Deep Belief Network. RBM parameters are estimated using an efficient training method that called Contrastive Divergence. Parallel implementation of RBM is available using different models such as OpenMP, and CUDA. But this project has been the first attempt to apply OpenACC model on RBM.

  15. Travelling wave solutions to nonlinear physical models by means of ...

    Indian Academy of Sciences (India)

    On the other hand, considerable attention has been given to problem of finding spe- cial types of analytic solutions to understand biological, physical and chemical phenomena modelled by NPDEs. Among the possible solutions, certain solutions may depend only on a single combination of variables such as travelling wave ...

  16. General classical solutions in the noncommutative CPN-1 model

    International Nuclear Information System (INIS)

    Foda, O.; Jack, I.; Jones, D.R.T.

    2002-01-01

    We give an explicit construction of general classical solutions for the noncommutative CP N-1 model in two dimensions, showing that they correspond to integer values for the action and topological charge. We also give explicit solutions for the Dirac equation in the background of these general solutions and show that the index theorem is satisfied

  17. Finite-element time-domain algorithms for modeling linear Debye and Lorentz dielectric dispersions at low frequencies.

    Science.gov (United States)

    Stoykov, Nikolay S; Kuiken, Todd A; Lowery, Madeleine M; Taflove, Allen

    2003-09-01

    We present what we believe to be the first algorithms that use a simple scalar-potential formulation to model linear Debye and Lorentz dielectric dispersions at low frequencies in the context of finite-element time-domain (FETD) numerical solutions of electric potential. The new algorithms, which permit treatment of multiple-pole dielectric relaxations, are based on the auxiliary differential equation method and are unconditionally stable. We validate the algorithms by comparison with the results of a previously reported method based on the Fourier transform. The new algorithms should be useful in calculating the transient response of biological materials subject to impulsive excitation. Potential applications include FETD modeling of electromyography, functional electrical stimulation, defibrillation, and effects of lightning and impulsive electric shock.

  18. An Evolutionary Search Algorithm for Covariate Models in Population Pharmacokinetic Analysis.

    Science.gov (United States)

    Yamashita, Fumiyoshi; Fujita, Atsuto; Sasa, Yukako; Higuchi, Yuriko; Tsuda, Masahiro; Hashida, Mitsuru

    2017-09-01

    Building a covariate model is a crucial task in population pharmacokinetics. This study develops a novel method for automated covariate modeling based on gene expression programming (GEP), which not only enables covariate selection, but also the construction of nonpolynomial relationships between pharmacokinetic parameters and covariates. To apply GEP to the extended nonlinear least squares analysis, the parameter consolidation and initial parameter value estimation algorithms were further developed and implemented. The entire program was coded in Java. The performance of the developed covariate model was evaluated for the population pharmacokinetic data of tobramycin. In comparison with the established covariate model, goodness-of-fit of the measured data was greatly improved by using only 2 additional adjustable parameters. Ten test runs yielded the same solution. In conclusion, the systematic exploration method is a potentially powerful tool for prescreening covariate models in population pharmacokinetic analysis. Copyright © 2017 American Pharmacists Association®. Published by Elsevier Inc. All rights reserved.

  19. New analytical solution for pyle-popovich's peritoneal dialysis model

    Energy Technology Data Exchange (ETDEWEB)

    Hamada, Hiroyuki; Sakiyama, Ryoichi; Okamoto, Masahiro; Tojo, Kakuji [Kyushi Institute of Technology, Fukuoka (Japan); Yamashita, Akihiro [Shonan Institute of Technology, Kanagwa (Japan)

    1999-08-01

    Continuous Ambulatory Peritoneal Dialysis (CAPD) is one of the standard treatments for kidney disease patients. A washing solution, called dialysate, is put into the peritoneal cavity to remove waste products and excess amounts of water in CAPD. The dialysate is exchanged four to five times a day by the patient. However, it is not easy to prescribe CAPD therapy, which may have precluded popularization of CAPD therapy. Popovich et al. constructed a mathematical model (P-P model) that applies to the prescription of the treatment schedule. It requires, however, a number of iterative calculations to obtain a exact numerical solution because the model is a set of nonlinear simultaneous ordinary differential equations. In this paper, the authors derived a new approximated analytical solution by employing a time-discrete technique, assuming all the parameters to be constant within each piecewise period of time for the P-P model. We have also described an algorithm of a numerical calculation with the new solution for clinical use with another analytical solution (Vonesh's solution). The new analytical solution consists of a forward solution (FW solution). The new analytical solution consists of a forward solution (FW solution), that is the solution for the plasma and dialysate concentrations from t{sub i} to t{sub i+1}(t{sub i}solution (BW solution) from t{sub i} to t{sub i-1}(t{sub i-1}solutions. The new analytical solution show an excellent agreement with the exact numerical solution for entire dwelling time. Moreover, optimized parameters with the new analytical solution show much smaller discrepancy than those with Vonesh's solution. Although the proposed method requires a slightly longer calculation time than Vonesh's it can simulate concentrations in

  20. Electronic design automation of analog ICs combining gradient models with multi-objective evolutionary algorithms

    CERN Document Server

    Rocha, Frederico AE; Lourenço, Nuno CC; Horta, Nuno CG

    2013-01-01

    This book applies to the scientific area of electronic design automation (EDA) and addresses the automatic sizing of analog integrated circuits (ICs). Particularly, this book presents an approach to enhance a state-of-the-art layout-aware circuit-level optimizer (GENOM-POF), by embedding statistical knowledge from an automatically generated gradient model into the multi-objective multi-constraint optimization kernel based on the NSGA-II algorithm. The results showed allow the designer to explore the different trade-offs of the solution space, both through the achieved device sizes, or the resp

  1. Stochastic time-dependent vehicle routing problem: Mathematical models and ant colony algorithm

    Directory of Open Access Journals (Sweden)

    Zhengyu Duan

    2015-11-01

    Full Text Available This article addresses the stochastic time-dependent vehicle routing problem. Two mathematical models named robust optimal schedule time model and minimum expected schedule time model are proposed for stochastic time-dependent vehicle routing problem, which can guarantee delivery within the time windows of customers. The robust optimal schedule time model only requires the variation range of link travel time, which can be conveniently derived from historical traffic data. In addition, the robust optimal schedule time model based on robust optimization method can be converted into a time-dependent vehicle routing problem. Moreover, an ant colony optimization algorithm is designed to solve stochastic time-dependent vehicle routing problem. As the improvements in initial solution and transition probability, ant colony optimization algorithm has a good performance in convergence. Through computational instances and Monte Carlo simulation tests, robust optimal schedule time model is proved to be better than minimum expected schedule time model in computational efficiency and coping with the travel time fluctuations. Therefore, robust optimal schedule time model is applicable in real road network.

  2. New Algorithms for Computing the Time-to-Collision in Freeway Traffic Simulation Models

    Directory of Open Access Journals (Sweden)

    Jia Hou

    2014-01-01

    Full Text Available Ways to estimate the time-to-collision are explored. In the context of traffic simulation models, classical lane-based notions of vehicle location are relaxed and new, fast, and efficient algorithms are examined. With trajectory conflicts being the main focus, computational procedures are explored which use a two-dimensional coordinate system to track the vehicle trajectories and assess conflicts. Vector-based kinematic variables are used to support the calculations. Algorithms based on boxes, circles, and ellipses are considered. Their performance is evaluated in the context of computational complexity and solution time. Results from these analyses suggest promise for effective and efficient analyses. A combined computation process is found to be very effective.

  3. Toward a Mesoscale Model for the Dynamics of Polymer Solutions

    Energy Technology Data Exchange (ETDEWEB)

    Miller, G H; Trebotich, D

    2006-10-02

    To model entire microfluidic systems containing solvated polymers we argue that it is necessary to have a numerical stability constraint governed only by the advective CFL condition. Advancements in the treatment of Kramers bead-rod polymer models are presented to enable tightly-coupled fluid-particle algorithms in the context of system-level modeling.

  4. From Point Clouds to Architectural Models: Algorithms for Shape Reconstruction

    Science.gov (United States)

    Canciani, M.; Falcolini, C.; Saccone, M.; Spadafora, G.

    2013-02-01

    The use of terrestrial laser scanners in architectural survey applications has become more and more common. Row data complexity, as given by scanner restitution, leads to several problems about design and 3D-modelling starting from Point Clouds. In this context we present a study on architectural sections and mathematical algorithms for their shape reconstruction, according to known or definite geometrical rules, focusing on shapes of different complexity. Each step of the semi-automatic algorithm has been developed using Mathematica software and CAD, integrating both programs in order to reconstruct a geometrical CAD model of the object. Our study is motivated by the fact that, for architectural survey, most of three dimensional modelling procedures concerning point clouds produce superabundant, but often unnecessary, information and are also very expensive in terms of cpu time using more and more sophisticated hardware and software. On the contrary, it's important to simplify/decimate the point cloud in order to recognize a particular form out of some definite geometric/architectonic shapes. Such a process consists of several steps: first the definition of plane sections and characterization of their architecture; secondly the construction of a continuous plane curve depending on some parameters. In the third step we allow the selection on the curve of some nodal points with given specific characteristics (symmetry, tangency conditions, shadowing exclusion, corners, … ). The fourth and last step is the construction of a best shape defined by the comparison with an abacus of known geometrical elements, such as moulding profiles, leading to a precise architectonical section. The algorithms have been developed and tested in very different situations and are presented in a case study of complex geometries such as some mouldings profiles in the Church of San Carlo alle Quattro Fontane.

  5. Modeling the Swift Bat Trigger Algorithm with Machine Learning

    Science.gov (United States)

    Graff, Philip B.; Lien, Amy Y.; Baker, John G.; Sakamoto, Takanori

    2016-01-01

    To draw inferences about gamma-ray burst (GRB) source populations based on Swift observations, it is essential to understand the detection efficiency of the Swift burst alert telescope (BAT). This study considers the problem of modeling the Swift / BAT triggering algorithm for long GRBs, a computationally expensive procedure, and models it using machine learning algorithms. A large sample of simulated GRBs from Lien et al. is used to train various models: random forests, boosted decision trees (with AdaBoost), support vector machines, and artificial neural networks. The best models have accuracies of greater than or equal to 97 percent (less than or equal to 3 percent error), which is a significant improvement on a cut in GRB flux, which has an accuracy of 89.6 percent (10.4 percent error). These models are then used to measure the detection efficiency of Swift as a function of redshift z, which is used to perform Bayesian parameter estimation on the GRB rate distribution. We find a local GRB rate density of n (sub 0) approaching 0.48 (sup plus 0.41) (sub minus 0.23) per cubic gigaparsecs per year with power-law indices of n (sub 1) approaching 1.7 (sup plus 0.6) (sub minus 0.5) and n (sub 2) approaching minus 5.9 (sup plus 5.7) (sub minus 0.1) for GRBs above and below a break point of z (redshift) (sub 1) approaching 6.8 (sup plus 2.8) (sub minus 3.2). This methodology is able to improve upon earlier studies by more accurately modeling Swift detection and using this for fully Bayesian model fitting.

  6. Multi-cut solutions in Chern-Simons matrix models

    Science.gov (United States)

    Morita, Takeshi; Sugiyama, Kento

    2018-04-01

    We elaborate the Chern-Simons (CS) matrix models at large N. The saddle point equations of these matrix models have a curious structure which cannot be seen in the ordinary one matrix models. Thanks to this structure, an infinite number of multi-cut solutions exist in the CS matrix models. Particularly we exactly derive the two-cut solutions at finite 't Hooft coupling in the pure CS matrix model. In the ABJM matrix model, we argue that some of multi-cut solutions might be interpreted as a condensation of the D2-brane instantons.

  7. Numerical model updating technique for structures using firefly algorithm

    Science.gov (United States)

    Sai Kubair, K.; Mohan, S. C.

    2018-03-01

    Numerical model updating is a technique used for updating the existing experimental models for any structures related to civil, mechanical, automobiles, marine, aerospace engineering, etc. The basic concept behind this technique is updating the numerical models to closely match with experimental data obtained from real or prototype test structures. The present work involves the development of numerical model using MATLAB as a computational tool and with mathematical equations that define the experimental model. Firefly algorithm is used as an optimization tool in this study. In this updating process a response parameter of the structure has to be chosen, which helps to correlate the numerical model developed with the experimental results obtained. The variables for the updating can be either material or geometrical properties of the model or both. In this study, to verify the proposed technique, a cantilever beam is analyzed for its tip deflection and a space frame has been analyzed for its natural frequencies. Both the models are updated with their respective response values obtained from experimental results. The numerical results after updating show that there is a close relationship that can be brought between the experimental and the numerical models.

  8. Development of modelling algorithm of technological systems by statistical tests

    Science.gov (United States)

    Shemshura, E. A.; Otrokov, A. V.; Chernyh, V. G.

    2018-03-01

    The paper tackles the problem of economic assessment of design efficiency regarding various technological systems at the stage of their operation. The modelling algorithm of a technological system was performed using statistical tests and with account of the reliability index allows estimating the level of machinery technical excellence and defining the efficiency of design reliability against its performance. Economic feasibility of its application shall be determined on the basis of service quality of a technological system with further forecasting of volumes and the range of spare parts supply.

  9. Heterogeneous Agents Model with the Worst Out Algorithm

    Czech Academy of Sciences Publication Activity Database

    Vošvrda, Miloslav; Vácha, Lukáš

    I, č. 1 (2007), s. 54-66 ISSN 1802-4696 R&D Projects: GA MŠk(CZ) LC06075; GA ČR(CZ) GA402/06/0990 Grant - others:GA UK(CZ) 454/2004/A-EK/FSV Institutional research plan: CEZ:AV0Z10750506 Keywords : Efficient Markets Hypothesis * Fractal Market Hypothesis * agents' investment horizons * agents' trading strategies * technical trading rules * heterogeneous agent model with stochastic memory * Worst out Algorithm Subject RIV: AH - Economics

  10. Two-Stage Electricity Demand Modeling Using Machine Learning Algorithms

    Directory of Open Access Journals (Sweden)

    Krzysztof Gajowniczek

    2017-10-01

    Full Text Available Forecasting of electricity demand has become one of the most important areas of research in the electric power industry, as it is a critical component of cost-efficient power system management and planning. In this context, accurate and robust load forecasting is supposed to play a key role in reducing generation costs, and deals with the reliability of the power system. However, due to demand peaks in the power system, forecasts are inaccurate and prone to high numbers of errors. In this paper, our contributions comprise a proposed data-mining scheme for demand modeling through peak detection, as well as the use of this information to feed the forecasting system. For this purpose, we have taken a different approach from that of time series forecasting, representing it as a two-stage pattern recognition problem. We have developed a peak classification model followed by a forecasting model to estimate an aggregated demand volume. We have utilized a set of machine learning algorithms to benefit from both accurate detection of the peaks and precise forecasts, as applied to the Polish power system. The key finding is that the algorithms can detect 96.3% of electricity peaks (load value equal to or above the 99th percentile of the load distribution and deliver accurate forecasts, with mean absolute percentage error (MAPE of 3.10% and resistant mean absolute percentage error (r-MAPE of 2.70% for the 24 h forecasting horizon.

  11. Using genetic algorithms to calibrate a water quality model.

    Science.gov (United States)

    Liu, Shuming; Butler, David; Brazier, Richard; Heathwaite, Louise; Khu, Soon-Thiam

    2007-03-15

    With the increasing concern over the impact of diffuse pollution on water bodies, many diffuse pollution models have been developed in the last two decades. A common obstacle in using such models is how to determine the values of the model parameters. This is especially true when a model has a large number of parameters, which makes a full range of calibration expensive in terms of computing time. Compared with conventional optimisation approaches, soft computing techniques often have a faster convergence speed and are more efficient for global optimum searches. This paper presents an attempt to calibrate a diffuse pollution model using a genetic algorithm (GA). Designed to simulate the export of phosphorus from diffuse sources (agricultural land) and point sources (human), the Phosphorus Indicators Tool (PIT) version 1.1, on which this paper is based, consisted of 78 parameters. Previous studies have indicated the difficulty of full range model calibration due to the number of parameters involved. In this paper, a GA was employed to carry out the model calibration in which all parameters were involved. A sensitivity analysis was also performed to investigate the impact of operators in the GA on its effectiveness in optimum searching. The calibration yielded satisfactory results and required reasonable computing time. The application of the PIT model to the Windrush catchment with optimum parameter values was demonstrated. The annual P loss was predicted as 4.4 kg P/ha/yr, which showed a good fitness to the observed value.

  12. A new parallelization algorithm of ocean model with explicit scheme

    Science.gov (United States)

    Fu, X. D.

    2017-08-01

    This paper will focus on the parallelization of ocean model with explicit scheme which is one of the most commonly used schemes in the discretization of governing equation of ocean model. The characteristic of explicit schema is that calculation is simple, and that the value of the given grid point of ocean model depends on the grid point at the previous time step, which means that one doesn’t need to solve sparse linear equations in the process of solving the governing equation of the ocean model. Aiming at characteristics of the explicit scheme, this paper designs a parallel algorithm named halo cells update with tiny modification of original ocean model and little change of space step and time step of the original ocean model, which can parallelize ocean model by designing transmission module between sub-domains. This paper takes the GRGO for an example to implement the parallelization of GRGO (Global Reduced Gravity Ocean model) with halo update. The result demonstrates that the higher speedup can be achieved at different problem size.

  13. Optimizing ion channel models using a parallel genetic algorithm on graphical processors.

    Science.gov (United States)

    Ben-Shalom, Roy; Aviv, Amit; Razon, Benjamin; Korngreen, Alon

    2012-01-01

    We have recently shown that we can semi-automatically constrain models of voltage-gated ion channels by combining a stochastic search algorithm with ionic currents measured using multiple voltage-clamp protocols. Although numerically successful, this approach is highly demanding computationally, with optimization on a high performance Linux cluster typically lasting several days. To solve this computational bottleneck we converted our optimization algorithm for work on a graphical processing unit (GPU) using NVIDIA's CUDA. Parallelizing the process on a Fermi graphic computing engine from NVIDIA increased the speed ∼180 times over an application running on an 80 node Linux cluster, considerably reducing simulation times. This application allows users to optimize models for ion channel kinetics on a single, inexpensive, desktop "super computer," greatly reducing the time and cost of building models relevant to neuronal physiology. We also demonstrate that the point of algorithm parallelization is crucial to its performance. We substantially reduced computing time by solving the ODEs (Ordinary Differential Equations) so as to massively reduce memory transfers to and from the GPU. This approach may be applied to speed up other data intensive applications requiring iterative solutions of ODEs. Copyright © 2012 Elsevier B.V. All rights reserved.

  14. An efficient algorithm for computation of solitary wave solutions to nonlinear differential equations

    Science.gov (United States)

    Ayub, Kamran; Khan, M. Yaqub; Mahmood-Ul-Hassan, Qazi; Ahmad, Jamshad

    2017-09-01

    Nonlinear mathematical problems and their solutions attain much attention in solitary waves. In soliton theory, an efficient tool to attain various types of soliton solutions is the \\exp (-φ (ζ ))-expansion technique. This article is devoted to find exact travelling wave solutions of Drinfeld-Sokolov equation via a reliable mathematical technique. By using the proposed technique, we attain soliton wave solution of various types. It is observed that the technique under discussion is user friendly with minimum computational work, and can be extended for physical problems of different nature in mathematical physics.

  15. Mathematical models and heuristic solutions for container positioning problems in port terminals

    DEFF Research Database (Denmark)

    Kallehauge, Louise Sibbesen

    2008-01-01

    , and conclusive remarks. The introduction provides a basis for appreciating the presented work and sets out the scope, motivation, purpose, and contributions of the thesis. Furthermore, the CPP is defined and described, an overview of port container terminal issues in general is provided, and relevant literature...... mathematical programming formulations. First, a conceptual mixed integer linear programming (MIP) model for the entire port container terminal is presented. Subsequently, two models for the CPP are suggested: A MIP model and a binary integer linear programming (BIP) model. The models provide a basis...... in parameters and input data. Two improvement routines are presented, one imbedded in the basic heuristic and the other constituting a repair algorithm with the purpose of improving an initial heuristic solution. The heuristic algorithm performance and a wide range of different planning strategies...

  16. Biosorption of chromium (VI) from aqueous solutions and ANN modelling.

    Science.gov (United States)

    Nag, Soma; Mondal, Abhijit; Bar, Nirjhar; Das, Sudip Kumar

    2017-08-01

    The use of sustainable, green and biodegradable natural wastes for Cr(VI) detoxification from the contaminated wastewater is considered as a challenging issue. The present research is aimed to assess the effectiveness of seven different natural biomaterials, such as jackfruit leaf, mango leaf, onion peel, garlic peel, bamboo leaf, acid treated rubber leaf and coconut shell powder, for Cr(VI) eradication from aqueous solution by biosorption process. Characterizations were conducted using SEM, BET and FTIR spectroscopy. The effects of operating parameters, viz., pH, initial Cr(VI) ion concentration, adsorbent dosages, contact time and temperature on metal removal efficiency, were studied. The biosorption mechanism was described by the pseudo-second-order model and Langmuir isotherm model. The biosorption process was exothermic, spontaneous and chemical (except garlic peel) in nature. The sequence of adsorption capacity was mango leaf > jackfruit leaf > acid treated rubber leaf > onion peel > bamboo leaf > garlic peel > coconut shell with maximum Langmuir adsorption capacity of 35.7 mg g -1 for mango leaf. The treated effluent can be reused. Desorption study suggested effective reuse of the adsorbents up to three cycles, and safe disposal method of the used adsorbents suggested biodegradability and sustainability of the process by reapplication of the spent adsorbent and ultimately leading towards zero wastages. The performances of the adsorbents were verified with wastewater from electroplating industry. The scale-up study reported for industrial applications. ANN modelling using multilayer perception with gradient descent (GD) and Levenberg-Marquart (LM) algorithm had been successfully used for prediction of Cr(VI) removal efficiency. The study explores the undiscovered potential of the natural waste materials for sustainable existence of small and medium sector industries, especially in the third world countries by protecting the environment by eco-innovation.

  17. A sequential quadratic programming algorithm using an incomplete solution of the subproblem

    Energy Technology Data Exchange (ETDEWEB)

    Murray, W. [Stanford Univ., CA (United States). Systems Optimization Lab.; Prieto, F.J. [Universidad `Carlos III` de Madrid (Spain). Dept. de Estadistica y Econometria

    1993-05-01

    We analyze sequential quadratic programming (SQP) methods to solve nonlinear constrained optimization problems that are more flexible in their definition than standard SQP methods. The type of flexibility introduced is motivated by the necessity to deviate from the standard approach when solving large problems. Specifically we no longer require a minimizer of the QP subproblem to be determined or particular Lagrange multiplier estimates to be used. Our main focus is on an SQP algorithm that uses a particular augmented Lagrangian merit function. New results are derived for this algorithm under weaker conditions than previously assumed; in particular, it is not assumed that the iterates lie on a compact set.

  18. A MATLAB GUI based algorithm for modelling Magnetotelluric data

    Science.gov (United States)

    Timur, Emre; Onsen, Funda

    2016-04-01

    The magnetotelluric method is an electromagnetic survey technique that images the electrical resistivity distribution of layers in subsurface depths. Magnetotelluric method measures simultaneously total electromagnetic field components such as both time-varying magnetic field B(t) and induced electric field E(t). At the same time, forward modeling of magnetotelluric method is so beneficial for survey planning purpose, for comprehending the method, especially for students, and as part of an iteration process in inverting measured data. The MTINV program can be used to model and to interpret geophysical electromagnetic (EM) magnetotelluric (MT) measurements using a horizontally layered earth model. This program uses either the apparent resistivity and phase components of the MT data together or the apparent resistivity data alone. Parameter optimization, which is based on linearized inversion method, can be utilized in 1D interpretations. In this study, a new MATLAB GUI based algorithm has been written for the 1D-forward modeling of magnetotelluric response function for multiple layers to use in educational studies. The code also includes an automatic Gaussian noise option for a demanded ratio value. Numerous applications were carried out and presented for 2,3 and 4 layer models and obtained theoretical data were interpreted using MTINV, in order to evaluate the initial parameters and effect of noise. Keywords: Education, Forward Modelling, Inverse Modelling, Magnetotelluric

  19. "Updates to Model Algorithms & Inputs for the Biogenic ...

    Science.gov (United States)

    We have developed new canopy emission algorithms and land use data for BEIS. Simulations with BEIS v3.4 and these updates in CMAQ v5.0.2 are compared these changes to the Model of Emissions of Gases and Aerosols from Nature (MEGAN) and evaluated the simulations against observations. This has resulted in improvements in model evaluations of modeled isoprene, NOx, and O3. The National Exposure Research Laboratory (NERL) Atmospheric Modeling and Analysis Division (AMAD) conducts research in support of EPA mission to protect human health and the environment. AMAD research program is engaged in developing and evaluating predictive atmospheric models on all spatial and temporal scales for forecasting the air quality and for assessing changes in air quality and air pollutant exposures, as affected by changes in ecosystem management and regulatory decisions. AMAD is responsible for providing a sound scientific and technical basis for regulatory policies based on air quality models to improve ambient air quality. The models developed by AMAD are being used by EPA, NOAA, and the air pollution community in understanding and forecasting not only the magnitude of the air pollution problem, but also in developing emission control policies and regulations for air quality improvements.

  20. Epidemic Modelling by Ripple-Spreading Network and Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Jian-Qin Liao

    2013-01-01

    Full Text Available Mathematical analysis and modelling is central to infectious disease epidemiology. This paper, inspired by the natural ripple-spreading phenomenon, proposes a novel ripple-spreading network model for the study of infectious disease transmission. The new epidemic model naturally has good potential for capturing many spatial and temporal features observed in the outbreak of plagues. In particular, using a stochastic ripple-spreading process simulates the effect of random contacts and movements of individuals on the probability of infection well, which is usually a challenging issue in epidemic modeling. Some ripple-spreading related parameters such as threshold and amplifying factor of nodes are ideal to describe the importance of individuals’ physical fitness and immunity. The new model is rich in parameters to incorporate many real factors such as public health service and policies, and it is highly flexible to modifications. A genetic algorithm is used to tune the parameters of the model by referring to historic data of an epidemic. The well-tuned model can then be used for analyzing and forecasting purposes. The effectiveness of the proposed method is illustrated by simulation results.

  1. Fast Time and Space Parallel Algorithms for Solution of Parabolic Partial Differential Equations

    Science.gov (United States)

    Fijany, Amir

    1993-01-01

    In this paper, fast time- and Space -Parallel agorithms for solution of linear parabolic PDEs are developed. It is shown that the seemingly strictly serial iterations of the time-stepping procedure for solution of the problem can be completed decoupled.

  2. Multiobjecitve Sampling Design for Calibration of Water Distribution Network Model Using Genetic Algorithm and Neural Network

    Directory of Open Access Journals (Sweden)

    Kourosh Behzadian

    2008-03-01

    Full Text Available In this paper, a novel multiobjective optimization model is presented for selecting optimal locations in the water distribution network (WDN with the aim of installing pressure loggers. The pressure data collected at optimal locations will be used later on in the calibration of the proposed WDN model. Objective functions consist of maximization of calibrated model prediction accuracy and minimization of the total cost for sampling design. In order to decrease the model run time, an optimization model has been developed using multiobjective genetic algorithm and adaptive neural network (MOGA-ANN. Neural networks (NNs are initially trained after a number of initial GA generations and periodically retrained and updated after generation of a specified number of full model-analyzed solutions. Trained NNs are replaced with the fitness evaluation of some chromosomes within the GA progress. Using cache prevents objective function evaluation of repetitive chromosomes within GA. Optimal solutions are obtained through pareto-optimal front with respect to the two objective functions. Results show that jointing NNs in MOGA for approximating portions of chromosomes’ fitness in each generation leads to considerable savings in model run time and can be promising for reducing run-time in optimization models with significant computational effort.

  3. Geometric algorithms for electromagnetic modeling of large scale structures

    Science.gov (United States)

    Pingenot, James

    With the rapid increase in the speed and complexity of integrated circuit designs, 3D full wave and time domain simulation of chip, package, and board systems becomes more and more important for the engineering of modern designs. Much effort has been applied to the problem of electromagnetic (EM) simulation of such systems in recent years. Major advances in boundary element EM simulations have led to O(n log n) simulations using iterative methods and advanced Fast. Fourier Transform (FFT), Multi-Level Fast Multi-pole Methods (MLFMM), and low-rank matrix compression techniques. These advances have been augmented with an explosion of multi-core and distributed computing technologies, however, realization of the full scale of these capabilities has been hindered by cumbersome and inefficient geometric processing. Anecdotal evidence from industry suggests that users may spend around 80% of turn-around time manipulating the geometric model and mesh. This dissertation addresses this problem by developing fast and efficient data structures and algorithms for 3D modeling of chips, packages, and boards. The methods proposed here harness the regular, layered 2D nature of the models (often referred to as "2.5D") to optimize these systems for large geometries. First, an architecture is developed for efficient storage and manipulation of 2.5D models. The architecture gives special attention to native representation of structures across various input models and special issues particular to 3D modeling. The 2.5D structure is then used to optimize the mesh systems First, circuit/EM co-simulation techniques are extended to provide electrical connectivity between objects. This concept is used to connect independently meshed layers, allowing simple and efficient 2D mesh algorithms to be used in creating a 3D mesh. Here, adaptive meshing is used to ensure that the mesh accurately models the physical unknowns (current and charge). Utilizing the regularized nature of 2.5D objects and

  4. Using the fuzzy modeling for the retrieval algorithms

    International Nuclear Information System (INIS)

    Mohamed, A.H

    2010-01-01

    A rapid growth in number and size of images in databases and world wide web (www) has created a strong need for more efficient search and retrieval systems to exploit the benefits of this large amount of information. However, the collection of this information is now based on the image technology. One of the limitations of the current image analysis techniques necessitates that most image retrieval systems use some form of text description provided by the users as the basis to index and retrieve images. To overcome this problem, the proposed system introduces the using of fuzzy modeling to describe the image by using the linguistic ambiguities. Also, the proposed system can include vague or fuzzy terms in modeling the queries to match the image descriptions in the retrieval process. This can facilitate the indexing and retrieving process, increase their performance and decrease its computational time . Therefore, the proposed system can improve the performance of the traditional image retrieval algorithms.

  5. Computational Analysis of 3D Ising Model Using Metropolis Algorithms

    International Nuclear Information System (INIS)

    Sonsin, A F; Cortes, M R; Nunes, D R; Gomes, J V; Costa, R S

    2015-01-01

    We simulate the Ising Model with the Monte Carlo method and use the algorithms of Metropolis to update the distribution of spins. We found that, in the specific case of the three-dimensional Ising Model, methods of Metropolis are efficient. Studying the system near the point of phase transition, we observe that the magnetization goes to zero. In our simulations we analyzed the behavior of the magnetization and magnetic susceptibility to verify the phase transition in a paramagnetic to ferromagnetic material. The behavior of the magnetization and of the magnetic susceptibility as a function of the temperature suggest a phase transition around KT/J ≈ 4.5 and was evidenced the problem of finite size of the lattice to work with large lattice. (paper)

  6. A genetic algorithm solution to the design of slat-type shading system

    Energy Technology Data Exchange (ETDEWEB)

    Tsangrassoulis, A. [Architectural Department, School of Engineering, University of Thessaly, 38334 Volos (Greece); Bourdakis, V. [Department of Planning and Regional Development, School of Engineering, University of Thessaly, 38334 Volos (Greece); Geros, V.; Santamouris, M. [Group of Building Environmental Studies, University of Athens, Building Physics 5, 15784, Athens (Greece)

    2006-11-15

    This article presents a technique for the design of slat-type blinds based on the their relative light intensity distribution under a uniform light source. The technique offers a new approach using a genetic algorithm in order to evolve the design according to a set of parameters. (author)

  7. An Efficient Algorithm for EM Scattering from Anatomically Realistic Human Head Model Using Parallel CG-FFT Method

    Directory of Open Access Journals (Sweden)

    Lei Zhao

    2014-01-01

    Full Text Available An efficient algorithm is proposed to analyze the electromagnetic scattering problem from a high resolution head model with pixel data format. The algorithm is based on parallel technique and the conjugate gradient (CG method combined with the fast Fourier transform (FFT. Using the parallel CG-FFT method, the proposed algorithm is very efficient and can solve very electrically large-scale problems which cannot be solved using the conventional CG-FFT method in a personal computer. The accuracy of the proposed algorithm is verified by comparing numerical results with analytical Mie-series solutions for dielectric spheres. Numerical experiments have demonstrated that the proposed method has good performance on parallel efficiency.

  8. Numerical solution of continuous-time DSGE models under Poisson uncertainty

    DEFF Research Database (Denmark)

    Posch, Olaf; Trimborn, Timo

    We propose a simple and powerful method for determining the transition process in continuous-time DSGE models under Poisson uncertainty numerically. The idea is to transform the system of stochastic differential equations into a system of functional differential equations of the retarded type. We...... then use the Waveform Relaxation algorithm to provide a guess of the policy function and solve the resulting system of ordinary differential equations by standard methods and fix-point iteration. Analytical solutions are provided as a benchmark from which our numerical method can be used to explore broader...... classes of models. We illustrate the algorithm simulating both the stochastic neoclassical growth model and the Lucas model under Poisson uncertainty which is motivated by the Barro-Rietz rare disaster hypothesis. We find that, even for non-linear policy functions, the maximum (absolute) error is very...

  9. Dataflow-Driven Crowdsourcing: Relational Models and Algorithms

    Directory of Open Access Journals (Sweden)

    D. A. Ustalov

    2016-01-01

    Full Text Available Recently, microtask crowdsourcing has become a popular approach for addressing various data mining problems. Crowdsourcing workflows for approaching such problems are composed of several data processing stages which require consistent representation for making the work reproducible. This paper is devoted to the problem of reproducibility and formalization of the microtask crowdsourcing process. A computational model for microtask crowdsourcing based on an extended relational model and a dataflow computational model has been proposed. The proposed collaborative dataflow computational model is designed for processing the input data sources by executing annotation stages and automatic synchronization stages simultaneously. Data processing stages and connections between them are expressed by using collaborative computation workflows represented as loosely connected directed acyclic graphs. A synchronous algorithm for executing such workflows has been described. The computational model has been evaluated by applying it to two tasks from the computational linguistics field: concept lexicalization refining in electronic thesauri and establishing hierarchical relations between such concepts. The “Add–Remove–Confirm” procedure is designed for adding the missing lexemes to the concepts while removing the odd ones. The “Genus–Species–Match” procedure is designed for establishing “is-a” relations between the concepts provided with the corresponding word pairs. The experiments involving both volunteers from popular online social networks and paid workers from crowdsourcing marketplaces confirm applicability of these procedures for enhancing lexical resources. 

  10. A quasilinear model for solute transport under unsaturated flow

    International Nuclear Information System (INIS)

    Houseworth, J.E.; Leem, J.

    2009-01-01

    We developed an analytical solution for solute transport under steady-state, two-dimensional, unsaturated flow and transport conditions for the investigation of high-level radioactive waste disposal. The two-dimensional, unsaturated flow problem is treated using the quasilinear flow method for a system with homogeneous material properties. Dispersion is modeled as isotropic and is proportional to the effective hydraulic conductivity. This leads to a quasilinear form for the transport problem in terms of a scalar potential that is analogous to the Kirchhoff potential for quasilinear flow. The solutions for both flow and transport scalar potentials take the form of Fourier series. The particular solution given here is for two sources of flow, with one source containing a dissolved solute. The solution method may easily be extended, however, for any combination of flow and solute sources under steady-state conditions. The analytical results for multidimensional solute transport problems, which previously could only be solved numerically, also offer an additional way to benchmark numerical solutions. An analytical solution for two-dimensional, steady-state solute transport under unsaturated flow conditions is presented. A specific case with two sources is solved but may be generalized to any combination of sources. The analytical results complement numerical solutions, which were previously required to solve this class of problems.

  11. Solutions of ward's modified chiral model

    International Nuclear Information System (INIS)

    Ioannidou, T.; Zakrzewski, W.J.

    1997-01-01

    We discuss the adaptation of Uhlenbeck's method of solving the chiral model in 2 Euclidean dimensions to Ward's modified chiral model in (2+1) dimensions. We show that the method reduces the problem of solving the second-order partial differential equations for the chiral field to solving a sequence of first-order partial differential equations for time dependent projector valued fields

  12. Physics Based Model for Cryogenic Chilldown and Loading. Part I: Algorithm

    Science.gov (United States)

    Luchinsky, Dmitry G.; Smelyanskiy, Vadim N.; Brown, Barbara

    2014-01-01

    We report the progress in the development of the physics based model for cryogenic chilldown and loading. The chilldown and loading is model as fully separated non-equilibrium two-phase flow of cryogenic fluid thermally coupled to the pipe walls. The solution follow closely nearly-implicit and semi-implicit algorithms developed for autonomous control of thermal-hydraulic systems developed by Idaho National Laboratory. A special attention is paid to the treatment of instabilities. The model is applied to the analysis of chilldown in rapid loading system developed at NASA-Kennedy Space Center. The nontrivial characteristic feature of the analyzed chilldown regime is its active control by dump valves. The numerical predictions are in reasonable agreement with the experimental time traces. The obtained results pave the way to the development of autonomous loading operation on the ground and space.

  13. Modelling environmental dynamics. Advances in goematic solutions

    Energy Technology Data Exchange (ETDEWEB)

    Paegelow, Martin [Toulouse-2 Univ., 31 (France). GEODE UMR 5602 CNRS; Camacho Olmedo, Maria Teresa (eds.) [Granada Univ (Spain). Dpto. de Analisis Geografico Regional y Geografia Fisica

    2008-07-01

    Modelling environmental dynamics is critical to understanding and predicting the evolution of the environment in response to the large number of influences including urbanisation, climate change and deforestation. Simulation and modelling provide support for decision making in environmental management. The first chapter introduces terminology and provides an overview of methodological modelling approaches which may be applied to environmental and complex dynamics. Based on this introduction this book illustrates various models applied to a large variety of themes: deforestation in tropical regions, fire risk, natural reforestation in European mountains, agriculture, biodiversity, urbanism, climate change and land management for decision support, etc. These case studies, provided by a large international spectrum of researchers and presented in a uniform structure, focus particularly on methods and model validation so that this book is not only aimed at researchers and graduates but also at professionals. (orig.)

  14. The WITCH Model. Structure, Baseline, Solutions.

    Energy Technology Data Exchange (ETDEWEB)

    Bosetti, V.; Massetti, E.; Tavoni, M.

    2007-07-01

    WITCH - World Induced Technical Change Hybrid - is a regionally disaggregated hard link hybrid global model with a neoclassical optimal growth structure (top down) and an energy input detail (bottom up). The model endogenously accounts for technological change, both through learning curves affecting prices of new vintages of capital and through R and D investments. The model features the main economic and environmental policies in each world region as the outcome of a dynamic game. WITCH belongs to the class of Integrated Assessment Models as it possesses a climate module that feeds climate changes back into the economy. In this paper we provide a thorough discussion of the model structure and baseline projections. We report detailed information on the evolution of energy demand, technology and CO2 emissions. Finally, we explicitly quantifiy the role of free riding in determining the emissions scenarios. (auth)

  15. A Sequential Quadratic Programming Algorithm Using an Incomplete Solution of the Subproblem

    Science.gov (United States)

    1990-09-01

    MURRAY* AND FRANCISCO J. PRIETOt *Systems Optimization Laboratory Department of Operations Research Stanford University tDept. de Automitica, Ingenieria ...The search direction The algorithm described in Section 2 allows for considerable flexibility of design. We de - scribe here the specific choices made...and Sons, Chichester/New York/Brisbane/Toronto/Singapore). [Fra88] C. Fraley (1988), Software performance on nonlinear least-squares problems, SOL

  16. Regularity of solutions of a phase field model

    KAUST Repository

    Amler, Thomas

    2013-01-01

    Phase field models are widely-used for modelling phase transition processes such as solidification, freezing or CO2 sequestration. In this paper, a phase field model proposed by G. Caginalp is considered. The existence and uniqueness of solutions are proved in the case of nonsmooth initial data. Continuity of solutions with respect to time is established. In particular, it is shown that the governing initial boundary value problem can be considered as a dynamical system. © 2013 International Press.

  17. The Single Row Routing Problem Revisited: A Solution Based on Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    Albert Y. Zomaya

    2002-01-01

    Full Text Available With the advent of VLSI technology, circuits with more than one million transistors have been integrated onto a single chip. As the complexity of ICs grows, the time and money spent on designing the circuits become more important. A large, often dominant, part of the cost and time required to design an IC is consumed in the routing operation. The routing of carriers, such as in IC chips and printed circuit boards, is a classical problem in Computer Aided Design. With the complexity inherent in VLSI circuits, high performance routers are necessary. In this paper, a crucial step in the channel routing technique, the single row routing (SRR problem, is considered. First, we discuss the relevance of SRR in the context of the general routing problem. Secondly, we show that heuristic algorithms are far from solving the general problem. Next, we introduce evolutionary computation, and, in particular, genetic algorithms (GAs as a justifiable method in solving the SRR problem. Finally, an efficient O (nk complexity technique based on GAs heuristic is obtained to solve the general SRR problem containing n nodes. Experimental results show that the algorithm is faster and can often generate better results than many of the leading heuristics proposed in the literature.

  18. Global Solution of Atmospheric Circulation Models with Humidity Effect

    OpenAIRE

    Luo, Hong

    2014-01-01

    The atmospheric circulation models are deduced from the very complex atmospheric circulation models based on the actual background and meteorological data. The models are able to show features of atmospheric circulation and are easy to be studied. It is proved that existence of global solutions to atmospheric circulation models with the use of the $T$-weakly continuous operator.

  19. Original analytic solution of a half-bridge modelled as a statically indeterminate system

    Science.gov (United States)

    Oanta, Emil M.; Panait, Cornel; Raicu, Alexandra; Barhalescu, Mihaela

    2016-12-01

    The paper presents an original computer based analytical model of a half-bridge belonging to a circular settling tank. The primary unknown is computed using the force method, the coefficients of the canonical equation being calculated using either the discretization of the bending moment diagram in trapezoids, or using the relations specific to the polygons. A second algorithm based on the method of initial parameters is also presented. Analyzing the new solution we came to the conclusion that most of the computer code developed for other model may be reused. The results are useful to evaluate the behavior of the structure and to compare with the results of the finite element models.

  20. Bayesian reliability modeling and assessment solution for NC machine tools under small-sample data

    Science.gov (United States)

    Yang, Zhaojun; Kan, Yingnan; Chen, Fei; Xu, Binbin; Chen, Chuanhai; Yang, Chuangui

    2015-11-01

    Although Markov chain Monte Carlo(MCMC) algorithms are accurate, many factors may cause instability when they are utilized in reliability analysis; such instability makes these algorithms unsuitable for widespread engineering applications. Thus, a reliability modeling and assessment solution aimed at small-sample data of numerical control(NC) machine tools is proposed on the basis of Bayes theories. An expert-judgment process of fusing multi-source prior information is developed to obtain the Weibull parameters' prior distributions and reduce the subjective bias of usual expert-judgment methods. The grid approximation method is applied to two-parameter Weibull distribution to derive the formulas for the parameters' posterior distributions and solve the calculation difficulty of high-dimensional integration. The method is then applied to the real data of a type of NC machine tool to implement a reliability assessment and obtain the mean time between failures(MTBF). The relative error of the proposed method is 5.8020×10-4 compared with the MTBF obtained by the MCMC algorithm. This result indicates that the proposed method is as accurate as MCMC. The newly developed solution for reliability modeling and assessment of NC machine tools under small-sample data is easy, practical, and highly suitable for widespread application in the engineering field; in addition, the solution does not reduce accuracy.

  1. On the multimode quadratic vibronic coupling problem: An open-ended solution using a parallel Lanczos algorithm

    International Nuclear Information System (INIS)

    Schuurman, Michael S.; Young, Richard A.; Yarkony, David R.

    2008-01-01

    We introduce an algorithm to solve the secular equation that arises in the time independent multimode expansion of the quadratic vibronic coupling problem in parallel. The implementation can handle expansions of arbitrary length, with the open-ended character of the algorithm achieved through the use of fine grained parallelism to partition the trial vectors. The characteristics of the algorithm are discussed and its utility is illustrated by determining a model photoelectron spectrum of the ethoxy radical (C 2 H 5 O) using a vibronic expansion involving more than 1 billion vibronic basis states. This calculation also represents the first determination of a photoelectron spectrum obtained using a diabatic Hamiltonian obtained with a recently introduced ab initio surface reshaping procedure

  2. Analytic solution of a five-direction radiation transport model

    International Nuclear Information System (INIS)

    Cramer, S.N.

    1988-01-01

    In order to test certain spatial and angular dependent Monte Carlo biasing techniques, a one-dimensional, one energy, two-media, five-direction radiation transport model has been devised for which an analytic solution exists. Although this solution is too long to be conveniently expressed in an explicit form, it can be easily evaluated on the smallest of computers. This solution is discussed in this paper. 1 ref

  3. Toward Developing Genetic Algorithms to Aid in Critical Infrastructure Modeling

    Energy Technology Data Exchange (ETDEWEB)

    2007-05-01

    Today’s society relies upon an array of complex national and international infrastructure networks such as transportation, telecommunication, financial and energy. Understanding these interdependencies is necessary in order to protect our critical infrastructure. The Critical Infrastructure Modeling System, CIMS©, examines the interrelationships between infrastructure networks. CIMS© development is sponsored by the National Security Division at the Idaho National Laboratory (INL) in its ongoing mission for providing critical infrastructure protection and preparedness. A genetic algorithm (GA) is an optimization technique based on Darwin’s theory of evolution. A GA can be coupled with CIMS© to search for optimum ways to protect infrastructure assets. This includes identifying optimum assets to enforce or protect, testing the addition of or change to infrastructure before implementation, or finding the optimum response to an emergency for response planning. This paper describes the addition of a GA to infrastructure modeling for infrastructure planning. It first introduces the CIMS© infrastructure modeling software used as the modeling engine to support the GA. Next, the GA techniques and parameters are defined. Then a test scenario illustrates the integration with CIMS© and the preliminary results.

  4. An efficient algorithm for corona simulation with complex chemical models

    Science.gov (United States)

    Villa, Andrea; Barbieri, Luca; Gondola, Marco; Leon-Garzon, Andres R.; Malgesini, Roberto

    2017-05-01

    The simulation of cold plasma discharges is a leading field of applied sciences with many applications ranging from pollutant control to surface treatment. Many of these applications call for the development of novel numerical techniques to implement fully three-dimensional corona solvers that can utilize complex and physically detailed chemical databases. This is a challenging task since it multiplies the difficulties inherent to a three-dimensional approach by the complexity of databases comprising tens of chemical species and hundreds of reactions. In this paper a novel approach, capable of reducing significantly the computational burden, is developed. The proposed method is based on a proper time stepping algorithm capable of decomposing the original problem into simpler ones: each of them has then been tackled with either finite element, finite volume or ordinary differential equations solvers. This last solver deals with the chemical model and its efficient implementation is one of the main contributions of this work.

  5. Parameters extraction of the three diode model for the multi-crystalline solar cell/module using Moth-Flame Optimization Algorithm

    International Nuclear Information System (INIS)

    Allam, Dalia; Yousri, D.A.; Eteiba, M.B.

    2016-01-01

    Highlights: • More detailed models are proposed to emulate the multi-crystalline solar cell/module. • Moth-Flame Optimizer (MFO) is proposed for the parameter extraction process. • The performance of MFO technique is compared with the recent optimization algorithms. • MFO algorithm converges to the optimal solution more rapidly and more accurately. • MFO algorithm accomplished with three diode model achieves the most accurate model. - Abstract: As a result of the wide prevalence of using the multi-crystalline silicon solar cells, an accurate mathematical model for these cells has become an important issue. Therefore, a three diode model is proposed as a more precise model to meet the relatively complicated physical behavior of the multi-crystalline silicon solar cells. The performance of this model is compared to the performance of both the double diode and the modified double diode models of the same cell/module. Therefore, there is a persistent need to keep searching for a more accurate optimization algorithm to estimate the more complicated models’ parameters. Hence, a proper optimization algorithm which is called Moth-Flame Optimizer (MFO), is proposed as a new optimization algorithm for the parameter extraction process of the three tested models based on data measured at laboratory and other data reported at previous literature. To verify the performance of the suggested technique, its results are compared with the results of the most recent and powerful techniques in the literature such as Hybrid Evolutionary (DEIM) and Flower Pollination (FPA) algorithms. Furthermore, evaluation analysis is performed for the three algorithms of the selected models at different environmental conditions. The results show that, MFO algorithm achieves the least Root Mean Square Error (RMSE), Mean Bias Error (MBE), Absolute Error at the Maximum Power Point (AEMPP) and best Coefficient of Determination. In addition, MFO is reaching to the optimal solution with the

  6. Fundamental resource-allocating model in colleges and universities based on Immune Clone Algorithms

    Science.gov (United States)

    Ye, Mengdie

    2017-05-01

    In this thesis we will seek the combination of antibodies and antigens converted from the optimal course arrangement and make an analogy with Immune Clone Algorithms. According to the character of the Algorithms, we apply clone, clone gene and clone selection to arrange courses. Clone operator can combine evolutionary search and random search, global search and local search. By cloning and clone mutating candidate solutions, we can find the global optimal solution quickly.

  7. Solutions manual to accompany finite mathematics models and applications

    CERN Document Server

    Morris, Carla C

    2015-01-01

    A solutions manual to accompany Finite Mathematics: Models and Applications In order to emphasize the main concepts of each chapter, Finite Mathematics: Models and Applications features plentiful pedagogical elements throughout such as special exercises, end notes, hints, select solutions, biographies of key mathematicians, boxed key principles, a glossary of important terms and topics, and an overview of use of technology. The book encourages the modeling of linear programs and their solutions and uses common computer software programs such as LINDO. In addition to extensive chapters on pr

  8. Two New PRP Conjugate Gradient Algorithms for Minimization Optimization Models.

    Directory of Open Access Journals (Sweden)

    Gonglin Yuan

    Full Text Available Two new PRP conjugate Algorithms are proposed in this paper based on two modified PRP conjugate gradient methods: the first algorithm is proposed for solving unconstrained optimization problems, and the second algorithm is proposed for solving nonlinear equations. The first method contains two aspects of information: function value and gradient value. The two methods both possess some good properties, as follows: 1 βk ≥ 0 2 the search direction has the trust region property without the use of any line search method 3 the search direction has sufficient descent property without the use of any line search method. Under some suitable conditions, we establish the global convergence of the two algorithms. We conduct numerical experiments to evaluate our algorithms. The numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems and that the second algorithm is effective for solving large-scale nonlinear equations.

  9. A Non-Symmetrical Solution Applying a Genetic Algorithm with Natural Crossover for the Structural Optimization of Truss Structures

    Directory of Open Access Journals (Sweden)

    Alvarado-Cárdenas R.

    2012-07-01

    Full Text Available In this research it is proposed a genetic algorithm with “natural crossover” that was applied to a continuous-discrete representation in order to optimize truss structures. The objective is to reduce the weight by restraining node displacement and limiting the cross sections to use. The solutions are combined applying two types of crossovers to the same representation, thus allowing to effectively explore the search space. The results are validated by comparing those found herein against those found in current literature for the case of the design of a 70 m span bridge truss structure. Solutions obtained are lighter and with different topology. Additionally, a case study is proposed, a greenhouse roof truss structure, in order to generate an actual application that is built in a practical scale and it is loaded afterwards to verify its strength.

  10. C4: Exploring Multiple Solutions in Graphical Models by Cluster Sampling.

    Science.gov (United States)

    Porway, Jake; Zhu, Song-Chun

    2011-09-01

    This paper presents a novel Markov Chain Monte Carlo (MCMC) inference algorithm called C(4)--Clustering with Cooperative and Competitive Constraints--for computing multiple solutions from posterior probabilities defined on graphical models, including Markov random fields (MRF), conditional random fields (CRF), and hierarchical models. The graphs may have both positive and negative edges for cooperative and competitive constraints. C(4) is a probabilistic clustering algorithm in the spirit of Swendsen-Wang. By turning the positive edges on/off probabilistically, C(4) partitions the graph into a number of connected components (ccps) and each ccp is a coupled subsolution with nodes connected by positive edges. Then, by turning the negative edges on/off probabilistically, C(4) obtains composite ccps (called cccps) with competing ccps connected by negative edges. At each step, C(4) flips the labels of all nodes in a cccp so that nodes in each ccp keep the same label while different ccps are assigned different labels to observe both positive and negative constraints. Thus, the algorithm can jump between multiple competing solutions (or modes of the posterior probability) in a single or a few steps. It computes multiple distinct solutions to preserve the intrinsic ambiguities and avoids premature commitments to a single solution that may not be valid given later context. C(4) achieves a mixing rate faster than existing MCMC methods, such as various Gibbs samplers and Swendsen-Wang cuts. It is also more "dynamic" than common optimization methods such as ICM, LBP, and graph cuts. We demonstrate the C(4) algorithm in line drawing interpretation, scene labeling, and object recognition.

  11. Anisotropic static solutions in modelling highly compact bodies

    Indian Academy of Sciences (India)

    Einstein field equations for static anisotropic spheres are solved and exact interior solutions obtained. This paper extends earlier treatments to include anisotropic models which accommodate a wider variety of physically viable energy densities. Two classes of solutions are possible. The first class contains the limiting case ...

  12. Solute transport modelling with the variable temporally dependent ...

    Indian Academy of Sciences (India)

    Pintu Das

    2018-02-07

    Feb 7, 2018 ... Abstract. In this present study, analytical and numerical solutions are obtained for solute transport modelling in homogeneous semi-infinite porous medium. The dispersion coefficient is assumed to be initial dispersion and velocity is assumed to be temporally dependent with initial seepage velocity. Also ...

  13. A Linked Simulation-Optimization (LSO) Model for Conjunctive Irrigation Management using Clonal Selection Algorithm

    Science.gov (United States)

    Islam, Sirajul; Talukdar, Bipul

    2016-09-01

    A Linked Simulation-Optimization (LSO) model based on a Clonal Selection Algorithm (CSA) was formulated for application in conjunctive irrigation management. A series of measures were considered for reducing the computational burden associated with the LSO approach. Certain modifications were incurred to the formulated CSA, so as to decrease the number of function evaluations. In addition, a simple problem specific code for a two dimensional groundwater flow simulation model was developed. The flow model was further simplified by a novel approach of area reduction, in order to save computational time in simulation. The LSO model was applied in the irrigation command of the Pagladiya Dam Project in Assam, India. With a view to evaluate the performance of the CSA, a Genetic Algorithm (GA) was used as a comparison base. The results from the CSA compared well with those from the GA. In fact, the CSA was found to consume less computational time than the GA while converging to the optimal solution, due to the modifications incurred in it.

  14. Trust-region based return mapping algorithm for implicit integration of elastic-plastic constitutive models

    Energy Technology Data Exchange (ETDEWEB)

    Lester, Brian [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Scherzinger, William [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-01-19

    Here, a new method for the solution of the non-linear equations forming the core of constitutive model integration is proposed. Specifically, the trust-region method that has been developed in the numerical optimization community is successfully modified for use in implicit integration of elastic-plastic models. Although attention here is restricted to these rate-independent formulations, the proposed approach holds substantial promise for adoption with models incorporating complex physics, multiple inelastic mechanisms, and/or multiphysics. As a first step, the non-quadratic Hosford yield surface is used as a representative case to investigate computationally challenging constitutive models. The theory and implementation are presented, discussed, and compared to other common integration schemes. Multiple boundary value problems are studied and used to verify the proposed algorithm and demonstrate the capabilities of this approach over more common methodologies. Robustness and speed are then investigated and compared to existing algorithms. Through these efforts, it is shown that the utilization of a trust-region approach leads to superior performance versus a traditional closest-point projection Newton-Raphson method and comparable speed and robustness to a line search augmented scheme.

  15. Trust-region based return mapping algorithm for implicit integration of elastic-plastic constitutive models

    Energy Technology Data Exchange (ETDEWEB)

    Lester, Brian T. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Scherzinger, William M. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-01-19

    A new method for the solution of the non-linear equations forming the core of constitutive model integration is proposed. Specifically, the trust-region method that has been developed in the numerical optimization community is successfully modified for use in implicit integration of elastic-plastic models. Although attention here is restricted to these rate-independent formulations, the proposed approach holds substantial promise for adoption with models incorporating complex physics, multiple inelastic mechanisms, and/or multiphysics. As a first step, the non-quadratic Hosford yield surface is used as a representative case to investigate computationally challenging constitutive models. The theory and implementation are presented, discussed, and compared to other common integration schemes. Multiple boundary value problems are studied and used to verify the proposed algorithm and demonstrate the capabilities of this approach over more common methodologies. Robustness and speed are then investigated and compared to existing algorithms. As a result through these efforts, it is shown that the utilization of a trust-region approach leads to superior performance versus a traditional closest-point projection Newton-Raphson method and comparable speed and robustness to a line search augmented scheme.

  16. Spatial multiobjective optimization of agricultural conservation practices using a SWAT model and an evolutionary algorithm.

    Science.gov (United States)

    Rabotyagov, Sergey; Campbell, Todd; Valcu, Adriana; Gassman, Philip; Jha, Manoj; Schilling, Keith; Wolter, Calvin; Kling, Catherine

    2012-12-09

    Finding the cost-efficient (i.e., lowest-cost) ways of targeting conservation practice investments for the achievement of specific water quality goals across the landscape is of primary importance in watershed management. Traditional economics methods of finding the lowest-cost solution in the watershed context (e.g.,(5,12,20)) assume that off-site impacts can be accurately described as a proportion of on-site pollution generated. Such approaches are unlikely to be representative of the actual pollution process in a watershed, where the impacts of polluting sources are often determined by complex biophysical processes. The use of modern physically-based, spatially distributed hydrologic simulation models allows for a greater degree of realism in terms of process representation but requires a development of a simulation-optimization framework where the model becomes an integral part of optimization. Evolutionary algorithms appear to be a particularly useful optimization tool, able to deal with the combinatorial nature of a watershed simulation-optimization problem and allowing the use of the full water quality model. Evolutionary algorithms treat a particular spatial allocation of conservation practices in a watershed as a candidate solution and utilize sets (populations) of candidate solutions iteratively applying stochastic operators of selection, recombination, and mutation to find improvements with respect to the optimization objectives. The optimization objectives in this case are to minimize nonpoint-source pollution in the watershed, simultaneously minimizing the cost of conservation practices. A recent and expanding set of research is attempting to use similar methods and integrates water quality models with broadly defined evolutionary optimization methods(3,4,9,10,13-15,17-19,22,23,25). In this application, we demonstrate a program which follows Rabotyagov et al.'s approach and integrates a modern and commonly used SWAT water quality model(7) with a

  17. Calibration of Uncertainty Analysis of the SWAT Model Using Genetic Algorithms and Bayesian Model Averaging

    Science.gov (United States)

    In this paper, the Genetic Algorithms (GA) and Bayesian model averaging (BMA) were combined to simultaneously conduct calibration and uncertainty analysis for the Soil and Water Assessment Tool (SWAT). In this hybrid method, several SWAT models with different structures are first selected; next GA i...

  18. A genetic algorithm solution for the operation of green LTE networks with energy and environment considerations

    KAUST Repository

    Ghazzai, Hakim

    2012-01-01

    The Base Station (BS) sleeping strategy has become a well-known technique to achieve energy savings in cellular networks by switching off redundant BSs mainly for lightly loaded networks. Besides, the exploitation of renewable energies, as additional power sources in smart grids, becomes a real challenge to network operators to reduce power costs. In this paper, we propose a method based on genetic algorithms that decreases the energy consumption of a Long-Term Evolution (LTE) cellular network by not only shutting down underutilized BSs but also by optimizing the amounts of energy procured from the smart grid without affecting the desired Quality of Service. © 2012 Springer-Verlag.

  19. Problèmes de type obstacle : algorithmes de résolution

    African Journals Online (AJOL)

    discrétisation de l'inéquation variationnelle par la méthode des éléments finis linéaires conduit à l'étude de trois algorithmes ... (mécanique, physique, économie, contrôle optimal .... pour tout point P de h. ∑ . h. K est un convexe fermé de 0h. H [4]. L'analogue discret du problème (2) est donnée par la formulation suivante :.

  20. Manufactured analytical solutions for isothermal full-Stokes ice sheet models

    Directory of Open Access Journals (Sweden)

    A. Sargent

    2010-08-01

    Full Text Available We present the detailed construction of a manufactured analytical solution to time-dependent and steady-state isothermal full-Stokes ice sheet problems. The solutions are constructed for two-dimensional flowline and three-dimensional full-Stokes ice sheet models with variable viscosity. The construction is done by choosing for the specified ice surface and bed a velocity distribution that satisfies both mass conservation and the kinematic boundary conditions. Then a compensatory stress term in the conservation of momentum equations and their boundary conditions is calculated to make the chosen velocity distributions as well as the chosen pressure field into exact solutions. By substituting different ice surface and bed geometry formulas into the derived solution formulas, analytical solutions for different geometries can be constructed.

    The boundary conditions can be specified as essential Dirichlet conditions or as periodic boundary conditions. By changing a parameter value, the analytical solutions allow investigation of algorithms for a different range of aspect ratios as well as for different, frozen or sliding, basal conditions. The analytical solutions can also be used to estimate the numerical error of the method in the case when the effects of the boundary conditions are eliminated, that is, when the exact solution values are specified as inflow and outflow boundary conditions.

  1. A solution algorithm for calculating photon radiation fields with the aid of the Monte Carlo method

    International Nuclear Information System (INIS)

    Zappe, D.

    1978-04-01

    The MCTEST program and its subroutines for the solution of the Boltzmann transport equation is presented. The program renders possible to calculate photon radiation fields of point or plane gamma sources. After changing two subroutines the calculation can also be carried out for the case of directed incidence of radiation on plane shields of iron or concrete. (author)

  2. Model-based optimization strategy of chiller driven liquid desiccant dehumidifier with genetic algorithm

    International Nuclear Information System (INIS)

    Wang, Xinli; Cai, Wenjian; Lu, Jiangang; Sun, Youxian; Zhao, Lei

    2015-01-01

    This study presents a model-based optimization strategy for an actual chiller driven dehumidifier of liquid desiccant dehumidification system operating with lithium chloride solution. By analyzing the characteristics of the components, energy predictive models for the components in the dehumidifier are developed. To minimize the energy usage while maintaining the outlet air conditions at the pre-specified set-points, an optimization problem is formulated with an objective function, the constraints of mechanical limitations and components interactions. Model-based optimization strategy using genetic algorithm is proposed to obtain the optimal set-points for desiccant solution temperature and flow rate, to minimize the energy usage in the dehumidifier. Experimental studies on an actual system are carried out to compare energy consumption between the proposed optimization and the conventional strategies. The results demonstrate that energy consumption using the proposed optimization strategy can be reduced by 12.2% in the dehumidifier operation. - Highlights: • Present a model-based optimization strategy for energy saving in LDDS. • Energy predictive models for components in dehumidifier are developed. • The Optimization strategy are applied and tested in an actual LDDS. • Optimization strategy can achieve energy savings by 12% during operation

  3. On the Dagan Model of solute transport in groundwater: Application to the Borden Site

    Science.gov (United States)

    Barry, D. A.; Coves, J.; Sposito, Garrison

    1988-10-01

    In the first part of this study, Sposito and Barry (1987) derived an ensemble-mean convection-dispersion equation (CDE) for tracer solute transport subject to a random velocity field. It was shown that the model dispersion coefficients originally presented by Dagan (1984) could be derived from the general expression for the dispersion coefficients in this mean CDE. Under the assumption of ergodicity, the Dagan model is used in this paper to predict chloride and bromide concentrations in the well-documented Borden aquifer experiment reported by Roberts and Mackay (1986). Because of a possible influence on the solute from the upper aquifer boundary, it was appropriate to apply the two-dimensional form of the model. A number of steps was necessary to reduce the three-dimensional raw data to a two-dimensional form, the main ones being integration over the vertical axis and the use of a gridding algorithm to form a two-dimensional solute concentration surface. Incomplete sampling of the solute plume during the early sampling sessions, as well as the assumptions made with respect to the data analysis, produce a rather large degree of uncertainty in the specification of the initial solute plume. These factors hinder a thorough experimental evaluation of the Dagan model. Data from some of the later sampling sessions were more complete, however, and the model predictions appeared to agree well with the field concentration data, especially in the preasymptotic region for the longitudinal dispersion coefficient.

  4. Maximum Likelihood in a Generalized Linear Finite Mixture Model by Using the EM Algorithm

    NARCIS (Netherlands)

    Jansen, R.C.

    A generalized linear finite mixture model and an EM algorithm to fit the model to data are described. By this approach the finite mixture model is embedded within the general framework of generalized linear models (GLMs). Implementation of the proposed EM algorithm can be readily done in statistical

  5. Solution to automatic generation control problem using firefly algorithm optimized I(λ)D(µ) controller.

    Science.gov (United States)

    Debbarma, Sanjoy; Saikia, Lalit Chandra; Sinha, Nidul

    2014-03-01

    Present work focused on automatic generation control (AGC) of a three unequal area thermal systems considering reheat turbines and appropriate generation rate constraints (GRC). A fractional order (FO) controller named as I(λ)D(µ) controller based on crone approximation is proposed for the first time as an appropriate technique to solve the multi-area AGC problem in power systems. A recently developed metaheuristic algorithm known as firefly algorithm (FA) is used for the simultaneous optimization of the gains and other parameters such as order of integrator (λ) and differentiator (μ) of I(λ)D(µ) controller and governor speed regulation parameters (R). The dynamic responses corresponding to optimized I(λ)D(µ) controller gains, λ, μ, and R are compared with that of classical integer order (IO) controllers such as I, PI and PID controllers. Simulation results show that the proposed I(λ)D(µ) controller provides more improved dynamic responses and outperforms the IO based classical controllers. Further, sensitivity analysis confirms the robustness of the so optimized I(λ)D(µ) controller to wide changes in system loading conditions and size and position of SLP. Proposed controller is also found to have performed well as compared to IO based controllers when SLP takes place simultaneously in any two areas or all the areas. Robustness of the proposed I(λ)D(µ) controller is also tested against system parameter variations. © 2013 ISA. Published by Elsevier Ltd. All rights reserved.

  6. Modeling solutions to Tanzania's physician workforce challenge

    Directory of Open Access Journals (Sweden)

    Alex J. Goodell

    2016-06-01

    Full Text Available Background: There is a great need for physicians in Tanzania. In 2012, there were approximately 0.31 physicians per 10,000 individuals nationwide, with a lower ratio in the rural areas, where the majority of the population resides. In response, universities across Tanzania have greatly increased the enrollment of medical students. Yet evidence suggests high attrition of medical graduates to other professions and emigration from rural areas where they are most needed. Objective: To estimate the future number of physicians practicing in Tanzania and the potential impact of interventions to improve retention, we built a model that tracks medical students from enrollment through clinical practice, from 1990 to 2025. Design: We designed a Markov process with 92 potential states capturing the movement of 25,000 medical students and physicians from medical training through employment. Work possibilities included clinical practice (divided into rural or urban, public or private, non-clinical work, and emigration. We populated and calibrated the model using a national 2005/2006 physician mapping survey, as well as graduation records, graduate tracking surveys, and other available data. Results: The model projects massive losses to clinical practice between 2016 and 2025, especially in rural areas. Approximately 56% of all medical school students enrolled between 2011 and 2020 will not be practicing medicine in Tanzania in 2025. Even with these losses, the model forecasts an increase in the physician-to-population ratio to 1.4 per 10,000 by 2025. Increasing the absorption of recent graduates into the public sector and/or developing a rural training track would ameliorate physician attrition in the most underserved areas. Conclusions: Tanzania is making significant investments in the training of physicians. Without linking these doctors to employment and ensuring their retention, the majority of this investment in medical education will be jeopardized.

  7. The use of genetic algorithms to model protoplanetary discs

    Science.gov (United States)

    Hetem, Annibal; Gregorio-Hetem, Jane

    2007-12-01

    The protoplanetary discs of T Tauri and Herbig Ae/Be stars have previously been studied using geometric disc models to fit their spectral energy distribution (SED). The simulations provide a means to reproduce the signatures of various circumstellar structures, which are related to different levels of infrared excess. With the aim of improving our previous model, which assumed a simple flat-disc configuration, we adopt here a reprocessing flared-disc model that assumes hydrostatic, radiative equilibrium. We have developed a method to optimize the parameter estimation based on genetic algorithms (GAs). This paper describes the implementation of the new code, which has been applied to Herbig stars from the Pico dos Dias Survey catalogue, in order to illustrate the quality of the fitting for a variety of SED shapes. The star AB Aur was used as a test of the GA parameter estimation, and demonstrates that the new code reproduces successfully a canonical example of the flared-disc model. The GA method gives a good quality of fit, but the range of input parameters must be chosen with caution, as unrealistic disc parameters can be derived. It is confirmed that the flared-disc model fits the flattened SEDs typical of Herbig stars; however, embedded objects (increasing SED slope) and debris discs (steeply decreasing SED slope) are not well fitted with this configuration. Even considering the limitation of the derived parameters, the automatic process of SED fitting provides an interesting tool for the statistical analysis of the circumstellar luminosity of large samples of young stars.

  8. Using genetic algorithm and TOPSIS for Xinanjiang model calibration with a single procedure

    Science.gov (United States)

    Cheng, Chun-Tian; Zhao, Ming-Yan; Chau, K. W.; Wu, Xin-Yu

    2006-01-01

    Genetic Algorithm (GA) is globally oriented in searching and thus useful in optimizing multiobjective problems, especially where the objective functions are ill-defined. Conceptual rainfall-runoff models that aim at predicting streamflow from the knowledge of precipitation over a catchment have become a basic tool for flood forecasting. The parameter calibration of a conceptual model usually involves the multiple criteria for judging the performances of observed data. However, it is often difficult to derive all objective functions for the parameter calibration problem of a conceptual model. Thus, a new method to the multiple criteria parameter calibration problem, which combines GA with TOPSIS (technique for order performance by similarity to ideal solution) for Xinanjiang model, is presented. This study is an immediate further development of authors' previous research (Cheng, C.T., Ou, C.P., Chau, K.W., 2002. Combining a fuzzy optimal model with a genetic algorithm to solve multi-objective rainfall-runoff model calibration. Journal of Hydrology, 268, 72-86), whose obvious disadvantages are to split the whole procedure into two parts and to become difficult to integrally grasp the best behaviors of model during the calibration procedure. The current method integrates the two parts of Xinanjiang rainfall-runoff model calibration together, simplifying the procedures of model calibration and validation and easily demonstrated the intrinsic phenomenon of observed data in integrity. Comparison of results with two-step procedure shows that the current methodology gives similar results to the previous method, is also feasible and robust, but simpler and easier to apply in practice.

  9. Design and selection of load control strategies using a multiple objective model and evolutionary algorithms

    International Nuclear Information System (INIS)

    Gomes, Alvaro; Antunes, Carlos Henggeler; Martins, Antonio Gomes

    2005-01-01

    This paper aims at presenting a multiple objective model to evaluate the attractiveness of the use of demand resources (through load management control actions) by different stakeholders and in diverse structure scenarios in electricity systems. For the sake of model flexibility, the multiple (and conflicting) objective functions of technical, economical and quality of service nature are able to capture distinct market scenarios and operating entities that may be interested in promoting load management activities. The computation of compromise solutions is made by resorting to evolutionary algorithms, which are well suited to tackle multiobjective problems of combinatorial nature herein involving the identification and selection of control actions to be applied to groups of loads. (Author)

  10. A local time stepping algorithm for GPU-accelerated 2D shallow water models

    Science.gov (United States)

    Dazzi, Susanna; Vacondio, Renato; Dal Palù, Alessandro; Mignosa, Paolo

    2018-01-01

    In the simulation of flooding events, mesh refinement is often required to capture local bathymetric features and/or to detail areas of interest; however, if an explicit finite volume scheme is adopted, the presence of small cells in the domain can restrict the allowable time step due to the stability condition, thus reducing the computational efficiency. With the aim of overcoming this problem, the paper proposes the application of a Local Time Stepping (LTS) strategy to a GPU-accelerated 2D shallow water numerical model able to handle non-uniform structured meshes. The algorithm is specifically designed to exploit the computational capability of GPUs, minimizing the overheads associated with the LTS implementation. The results of theoretical and field-scale test cases show that the LTS model guarantees appreciable reductions in the execution time compared to the traditional Global Time Stepping strategy, without compromising the solution accuracy.

  11. The Support Reduction Algorithm for Computing Non-Parametric Function Estimates in Mixture Models

    OpenAIRE

    GROENEBOOM, PIET; JONGBLOED, GEURT; WELLNER, JON A.

    2008-01-01

    In this paper, we study an algorithm (which we call the support reduction algorithm) that can be used to compute non-parametric M-estimators in mixture models. The algorithm is compared with natural competitors in the context of convex regression and the ‘Aspect problem’ in quantum physics.

  12. A simple and efficient parallel FFT algorithm using the BSP model

    NARCIS (Netherlands)

    Bisseling, R.H.; Inda, M.A.

    2000-01-01

    In this paper we present a new parallel radix FFT algorithm based on the BSP model Our parallel algorithm uses the groupcyclic distribution family which makes it simple to understand and easy to implement We show how to reduce the com munication cost of the algorithm by a factor of three in the case

  13. Assimilation of low-level wind in a high-resolution mesoscale model using the back and forth nudging algorithm

    Directory of Open Access Journals (Sweden)

    Jean-François Mahfouf

    2012-06-01

    Full Text Available The performance of a new data assimilation algorithm called back and forth nudging (BFN is evaluated using a high-resolution numerical mesoscale model and simulated wind observations in the boundary layer. This new algorithm, of interest for the assimilation of high-frequency observations provided by ground-based active remote-sensing instruments, is straightforward to implement in a realistic atmospheric model. The convergence towards a steady-state profile can be achieved after five iterations of the BFN algorithm, and the algorithm provides an improved solution with respect to direct nudging. It is shown that the contribution of the nudging term does not dominate over other model physical and dynamical tendencies. Moreover, by running backward integrations with an adiabatic version of the model, the nudging coefficients do not need to be increased in order to stabilise the numerical equations. The ability of BFN to produce model changes upstream from the observations, in a similar way to 4-D-Var assimilation systems, is demonstrated. The capacity of the model to adjust to rapid changes in wind direction with the BFN is a first encouraging step, for example, to improve the detection and prediction of low-level wind shear phenomena through high-resolution mesoscale modelling over airports.

  14. An algorithm for analytical solution of basic problems featuring elastostatic bodies with cavities and surface flaws

    Science.gov (United States)

    Penkov, V. B.; Levina, L. V.; Novikova, O. S.; Shulmin, A. S.

    2018-03-01

    Herein we propose a methodology for structuring a full parametric analytical solution to problems featuring elastostatic media based on state-of-the-art computing facilities that support computerized algebra. The methodology includes: direct and reverse application of P-Theorem; methods of accounting for physical properties of media; accounting for variable geometrical parameters of bodies, parameters of boundary states, independent parameters of volume forces, and remote stress factors. An efficient tool to address the task is the sustainable method of boundary states originally designed for the purposes of computerized algebra and based on the isomorphism of Hilbertian spaces of internal states and boundary states of bodies. We performed full parametric solutions of basic problems featuring a ball with a nonconcentric spherical cavity, a ball with a near-surface flaw, and an unlimited medium with two spherical cavities.

  15. Recursive algorithm for arrays of generalized Bessel functions: Numerical access to Dirac-Volkov solutions.

    Science.gov (United States)

    Lötstedt, Erik; Jentschura, Ulrich D

    2009-02-01

    In the relativistic and the nonrelativistic theoretical treatment of moderate and high-power laser-matter interaction, the generalized Bessel function occurs naturally when a Schrödinger-Volkov and Dirac-Volkov solution is expanded into plane waves. For the evaluation of cross sections of quantum electrodynamic processes in a linearly polarized laser field, it is often necessary to evaluate large arrays of generalized Bessel functions, of arbitrary index but with fixed arguments. We show that the generalized Bessel function can be evaluated, in a numerically stable way, by utilizing a recurrence relation and a normalization condition only, without having to compute any initial value. We demonstrate the utility of the method by illustrating the quantum-classical correspondence of the Dirac-Volkov solutions via numerical calculations.

  16. A Stress Update Algorithm for Constitutive Models of Glassy Polymers

    Science.gov (United States)

    Danielsson, Mats

    2013-06-01

    A semi-implicit stress update algorithm is developed for the elastic-viscoplastic behavior of glassy polymers. The case of near rate-insensitivity is addressed, and the stress update algorithm is designed to handle this case robustly. A consistent tangent stiffness matrix is derived based on a full linearization of the internal virtual work. The stress update algorithm and (a slightly modified) tangent stiffness matrix are implemented in a commercial finite element program. The stress update algorithm is tested on a large boundary value problem for illustrative purposes.

  17. Spectral bisection algorithm for solving Schrodinger equation using upper and lower solutions

    Directory of Open Access Journals (Sweden)

    Qutaibeh Deeb Katatbeh

    2007-10-01

    Full Text Available This paper establishes a new criteria for obtaining a sequence of upper and lower bounds for the ground state eigenvalue of Schr"odinger equation $ -Deltapsi(r+V(rpsi(r=Epsi(r$ in $N$ spatial dimensions. Based on this proposed criteria, we prove a new comparison theorem in quantum mechanics for the ground state eigenfunctions of Schrodinger equation. We determine also lower and upper solutions for the exact wave function of the ground state eigenfunctions using the computed upper and lower bounds for the eigenvalues obtained by variational methods. In other words, by using this criteria, we prove that the substitution of the lower(upper bound of the eigenvalue in Schrodinger equation leads to an upper(lower solution. Finally, two proposed iteration approaches lead to an exact convergent sequence of solutions. The first one uses Raielgh-Ritz theorem. Meanwhile, the second approach uses a new numerical spectral bisection technique. We apply our results for a wide class of potentials in quantum mechanics such as sum of power-law potentials in quantum mechanics.

  18. Theoretical framework for a dynamic cone-beam reconstruction algorithm based on a dynamic particle model.

    Science.gov (United States)

    Grangeat, Pierre; Koenig, Anne; Rodet, Thomas; Bonnet, Stéphane

    2002-08-07

    Dynamic cone-beam reconstruction algorithms are required to reconstruct three-dimensional (3D) image sequences on dynamic 3D CT combining multi-row two-dimensional (2D) detectors and sub-second scanners. The speed-up of the rotating gantry allows one to improve the temporal resolution of the image sequence, but at the same time, it implies increase in the dose delivered during a given time period to keep constant the signal-to-noise ratio associated with each frame. The alternative solution proposed in this paper is to process data acquisition on several half-turns in order to reduce the dose delivered per rotation with the same signal-to-noise ratio. In order to compensate for time evolution and motion artefacts, we propose to use a dynamic particle model to describe the object evolution during the scan. In this article, we first introduce the dynamic particle model and the dynamic CT acquisition model. Then, we explain the principle of the proposed dynamic cone-beam reconstruction algorithm. Lastly, we present preliminary results on simulated data.

  19. Development of Web-Based Menu Planning Support System and its Solution Using Genetic Algorithm

    Science.gov (United States)

    Kashima, Tomoko; Matsumoto, Shimpei; Ishii, Hiroaki

    2009-10-01

    Recently lifestyle-related diseases have become an object of public concern, while at the same time people are being more health conscious. As an essential factor for causing the lifestyle-related diseases, we assume that the knowledge circulation on dietary habits is still insufficient. This paper focuses on everyday meals close to our life and proposes a well-balanced menu planning system as a preventive measure of lifestyle-related diseases. The system is developed by using a Web-based frontend and it provides multi-user services and menu information sharing capabilities like social networking services (SNS). The system is implemented on a Web server running Apache (HTTP server software), MySQL (database management system), and PHP (scripting language for dynamic Web pages). For the menu planning, a genetic algorithm is applied by understanding this problem as multidimensional 0-1 integer programming.

  20. Finding a pareto-optimal solution for multi-region models subject to capital trade and spillover externalities

    Energy Technology Data Exchange (ETDEWEB)

    Leimbach, Marian [Potsdam-Institut fuer Klimafolgenforschung e.V., Potsdam (Germany); Eisenack, Klaus [Oldenburg Univ. (Germany). Dept. of Economics and Statistics

    2008-11-15

    In this paper we present an algorithm that deals with trade interactions within a multi-region model. In contrast to traditional approaches this algorithm is able to handle spillover externalities. Technological spillovers are expected to foster the diffusion of new technologies, which helps to lower the cost of climate change mitigation. We focus on technological spillovers which are due to capital trade. The algorithm of finding a pareto-optimal solution in an intertemporal framework is embedded in a decomposed optimization process. The paper analyzes convergence and equilibrium properties of this algorithm. In the final part of the paper, we apply the algorithm to investigate possible impacts of technological spillovers. While benefits of technological spillovers are significant for the capital-importing region, benefits for the capital-exporting region depend on the type of regional disparities and the resulting specialization and terms-of-trade effects. (orig.)

  1. Genetic algorithm based optimization of advanced solar cell designs modeled in Silvaco AtlasTM

    OpenAIRE

    Utsler, James

    2006-01-01

    A genetic algorithm was used to optimize the power output of multi-junction solar cells. Solar cell operation was modeled using the Silvaco ATLASTM software. The output of the ATLASTM simulation runs served as the input to the genetic algorithm. The genetic algorithm was run as a diffusing computation on a network of eighteen dual processor nodes. Results showed that the genetic algorithm produced better power output optimizations when compared with the results obtained using the hill cli...

  2. A Cost-Effective Tracking Algorithm for Hypersonic Glide Vehicle Maneuver Based on Modified Aerodynamic Model

    Directory of Open Access Journals (Sweden)

    Yu Fan

    2016-10-01

    Full Text Available In order to defend the hypersonic glide vehicle (HGV, a cost-effective single-model tracking algorithm using Cubature Kalman filter (CKF is proposed in this paper based on modified aerodynamic model (MAM as process equation and radar measurement model as measurement equation. In the existing aerodynamic model, the two control variables attack angle and bank angle cannot be measured by the existing radar equipment and their control laws cannot be known by defenders. To establish the process equation, the MAM for HGV tracking is proposed by using additive white noise to model the rates of change of the two control variables. For the ease of comparison several multiple model algorithms based on CKF are presented, including interacting multiple model (IMM algorithm, adaptive grid interacting multiple model (AGIMM algorithm and hybrid grid multiple model (HGMM algorithm. The performances of these algorithms are compared and analyzed according to the simulation results. The simulation results indicate that the proposed tracking algorithm based on modified aerodynamic model has the best tracking performance with the best accuracy and least computational cost among all tracking algorithms in this paper. The proposed algorithm is cost-effective for HGV tracking.

  3. Introducing Elitist Black-Box Models: When Does Elitist Selection Weaken the Performance of Evolutionary Algorithms?

    OpenAIRE

    Doerr, Carola; Lengler, Johannes

    2015-01-01

    Black-box complexity theory provides lower bounds for the runtime of black-box optimizers like evolutionary algorithms and serves as an inspiration for the design of new genetic algorithms. Several black-box models covering different classes of algorithms exist, each highlighting a different aspect of the algorithms under considerations. In this work we add to the existing black-box notions a new \\emph{elitist black-box model}, in which algorithms are required to base all decisions solely on ...

  4. A multiobjective interval programming model for wind-hydrothermal power system dispatching using 2-step optimization algorithm.

    Science.gov (United States)

    Ren, Kun; Jihong, Qu

    2014-01-01

    Wind-hydrothermal power system dispatching has received intensive attention in recent years because it can help develop various reasonable plans to schedule the power generation efficiency. But future data such as wind power output and power load would not be accurately predicted and the nonlinear nature involved in the complex multiobjective scheduling model; therefore, to achieve accurate solution to such complex problem is a very difficult task. This paper presents an interval programming model with 2-step optimization algorithm to solve multiobjective dispatching. Initially, we represented the future data into interval numbers and simplified the object function to a linear programming problem to search the feasible and preliminary solutions to construct the Pareto set. Then the simulated annealing method was used to search the optimal solution of initial model. Thorough experimental results suggest that the proposed method performed reasonably well in terms of both operating efficiency and precision.

  5. A Multiobjective Interval Programming Model for Wind-Hydrothermal Power System Dispatching Using 2-Step Optimization Algorithm

    Directory of Open Access Journals (Sweden)

    Kun Ren

    2014-01-01

    Full Text Available Wind-hydrothermal power system dispatching has received intensive attention in recent years because it can help develop various reasonable plans to schedule the power generation efficiency. But future data such as wind power output and power load would not be accurately predicted and the nonlinear nature involved in the complex multiobjective scheduling model; therefore, to achieve accurate solution to such complex problem is a very difficult task. This paper presents an interval programming model with 2-step optimization algorithm to solve multiobjective dispatching. Initially, we represented the future data into interval numbers and simplified the object function to a linear programming problem to search the feasible and preliminary solutions to construct the Pareto set. Then the simulated annealing method was used to search the optimal solution of initial model. Thorough experimental results suggest that the proposed method performed reasonably well in terms of both operating efficiency and precision.

  6. A multiobjective optimization model and an orthogonal design-based hybrid heuristic algorithm for regional urban mining management problems.

    Science.gov (United States)

    Wu, Hao; Wan, Zhong

    2018-02-01

    In this paper, a multiobjective mixed-integer piecewise nonlinear programming model (MOMIPNLP) is built to formulate the management problem of urban mining system, where the decision variables are associated with buy-back pricing, choices of sites, transportation planning, and adjustment of production capacity. Different from the existing approaches, the social negative effect, generated from structural optimization of the recycling system, is minimized in our model, as well as the total recycling profit and utility from environmental improvement are jointly maximized. For solving the problem, the MOMIPNLP model is first transformed into an ordinary mixed-integer nonlinear programming model by variable substitution such that the piecewise feature of the model is removed. Then, based on technique of orthogonal design, a hybrid heuristic algorithm is developed to find an approximate Pareto-optimal solution, where genetic algorithm is used to optimize the structure of search neighborhood, and both local branching algorithm and relaxation-induced neighborhood search algorithm are employed to cut the searching branches and reduce the number of variables in each branch. Numerical experiments indicate that this algorithm spends less CPU (central processing unit) time in solving large-scale regional urban mining management problems, especially in comparison with the similar ones available in literature. By case study and sensitivity analysis, a number of practical managerial implications are revealed from the model. Since the metal stocks in society are reliable overground mineral sources, urban mining has been paid great attention as emerging strategic resources in an era of resource shortage. By mathematical modeling and development of efficient algorithms, this paper provides decision makers with useful suggestions on the optimal design of recycling system in urban mining. For example, this paper can answer how to encourage enterprises to join the recycling activities

  7. Numerical Algorithms for Deterministic Impulse Control Models with Applications

    NARCIS (Netherlands)

    Grass, D.; Chahim, M.

    2012-01-01

    Abstract: In this paper we describe three different algorithms, from which two (as far as we know) are new in the literature. We take both the size of the jump as the jump times as decision variables. The first (new) algorithm considers an Impulse Control problem as a (multipoint) Boundary Value

  8. Modeling Complex Chemical Systems: Problems and Solutions

    Science.gov (United States)

    van Dijk, Jan

    2016-09-01

    Non-equilibrium plasmas in complex gas mixtures are at the heart of numerous contemporary technologies. They typically contain dozens to hundreds of species, involved in hundreds to thousands of reactions. Chemists and physicists have always been interested in what are now called chemical reduction techniques (CRT's). The idea of such CRT's is that they reduce the number of species that need to be considered explicitly without compromising the validity of the model. This is usually achieved on the basis of an analysis of the reaction time scales of the system under study, which identifies species that are in partial equilibrium after a given time span. The first such CRT that has been widely used in plasma physics was developed in the 1960's and resulted in the concept of effective ionization and recombination rates. It was later generalized to systems in which multiple levels are effected by transport. In recent years there has been a renewed interest in tools for chemical reduction and reaction pathway analysis. An example of the latter is the PumpKin tool. Another trend is that techniques that have previously been developed in other fields of science are adapted as to be able to handle the plasma state of matter. Examples are the Intrinsic Low Dimension Manifold (ILDM) method and its derivatives, which originate from combustion engineering, and the general-purpose Principle Component Analysis (PCA) technique. In this contribution we will provide an overview of the most common reduction techniques, then critically assess the pros and cons of the methods that have gained most popularity in recent years. Examples will be provided for plasmas in argon and carbon dioxide.

  9. A deflation based parallel algorithm for spectral element solution of the incompressible Navier-Stokes equations

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, P.F. [Brown Univ., Providence, RI (United States)

    1996-12-31

    Efficient solution of the Navier-Stokes equations in complex domains is dependent upon the availability of fast solvers for sparse linear systems. For unsteady incompressible flows, the pressure operator is the leading contributor to stiffness, as the characteristic propagation speed is infinite. In the context of operator splitting formulations, it is the pressure solve which is the most computationally challenging, despite its elliptic origins. We seek to improve existing spectral element iterative methods for the pressure solve in order to overcome the slow convergence frequently observed in the presence of highly refined grids or high-aspect ratio elements.

  10. Hydrodynamics with strength: scaling-invariant solutions for elastic-plastic cavity expansion models

    Science.gov (United States)

    Albright, Jason; Ramsey, Scott; Baty, Roy

    2017-11-01

    Spherical cavity expansion (SCE) models are used to describe idealized detonation and high-velocity impact in a variety of materials. The common theme in SCE models is the presence of a pressure-driven cavity or void within a domain comprised of plastic and elastic response sub-regions. In past work, the yield criterion characterizing material strength in the plastic sub-region is usually taken for granted and assumed to take a known functional form restrictive to certain classes of materials, e.g. ductile metals or brittle geologic materials. Our objective is to systematically determine a general functional form for the yield criterion under the additional requirement that the SCE admits a similarity solution. Solutions determined under this additional requirement have immediate implications toward development of new compressible flow algorithm verification test problems. However, more importantly, these results also provide novel insight into modeling the yield criteria from the perspective of hydrodynamic scaling.

  11. Hybrid nested sampling algorithm for Bayesian model selection applied to inverse subsurface flow problems

    International Nuclear Information System (INIS)

    Elsheikh, Ahmed H.; Wheeler, Mary F.; Hoteit, Ibrahim

    2014-01-01

    A Hybrid Nested Sampling (HNS) algorithm is proposed for efficient Bayesian model calibration and prior model selection. The proposed algorithm combines, Nested Sampling (NS) algorithm, Hybrid Monte Carlo (HMC) sampling and gradient estimation using Stochastic Ensemble Method (SEM). NS is an efficient sampling algorithm that can be used for Bayesian calibration and estimating the Bayesian evidence for prior model selection. Nested sampling has the advantage of computational feasibility. Within the nested sampling algorithm, a constrained sampling step is performed. For this step, we utilize HMC to reduce the correlation between successive sampled states. HMC relies on the gradient of the logarithm of the posterior distribution, which we estimate using a stochastic ensemble method based on an ensemble of directional derivatives. SEM only requires forward model runs and the simulator is then used as a black box and no adjoint code is needed. The developed HNS algorithm is successfully applied for Bayesian calibration and prior model selection of several nonlinear subsurface flow problems

  12. Quadratic adaptive algorithm for solving cardiac action potential models.

    Science.gov (United States)

    Chen, Min-Hung; Chen, Po-Yuan; Luo, Ching-Hsing

    2016-10-01

    An adaptive integration method is proposed for computing cardiac action potential models accurately and efficiently. Time steps are adaptively chosen by solving a quadratic formula involving the first and second derivatives of the membrane action potential. To improve the numerical accuracy, we devise an extremum-locator (el) function to predict the local extremum when approaching the peak amplitude of the action potential. In addition, the time step restriction (tsr) technique is designed to limit the increase in time steps, and thus prevent the membrane potential from changing abruptly. The performance of the proposed method is tested using the Luo-Rudy phase 1 (LR1), dynamic (LR2), and human O'Hara-Rudy dynamic (ORd) ventricular action potential models, and the Courtemanche atrial model incorporating a Markov sodium channel model. Numerical experiments demonstrate that the action potential generated using the proposed method is more accurate than that using the traditional Hybrid method, especially near the peak region. The traditional Hybrid method may choose large time steps near to the peak region, and sometimes causes the action potential to become distorted. In contrast, the proposed new method chooses very fine time steps in the peak region, but large time steps in the smooth region, and the profiles are smoother and closer to the reference solution. In the test on the stiff Markov ionic channel model, the Hybrid blows up if the allowable time step is set to be greater than 0.1ms. In contrast, our method can adjust the time step size automatically, and is stable. Overall, the proposed method is more accurate than and as efficient as the traditional Hybrid method, especially for the human ORd model. The proposed method shows improvement for action potentials with a non-smooth morphology, and it needs further investigation to determine whether the method is helpful during propagation of the action potential. Copyright © 2016 Elsevier Ltd. All rights

  13. Algorithms for Bayesian network modeling and reliability assessment of infrastructure systems

    International Nuclear Information System (INIS)

    Tien, Iris; Der Kiureghian, Armen

    2016-01-01

    Novel algorithms are developed to enable the modeling of large, complex infrastructure systems as Bayesian networks (BNs). These include a compression algorithm that significantly reduces the memory storage required to construct the BN model, and an updating algorithm that performs inference on compressed matrices. These algorithms address one of the major obstacles to widespread use of BNs for system reliability assessment, namely the exponentially increasing amount of information that needs to be stored as the number of components in the system increases. The proposed compression and inference algorithms are described and applied to example systems to investigate their performance compared to that of existing algorithms. Orders of magnitude savings in memory storage requirement are demonstrated using the new algorithms, enabling BN modeling and reliability analysis of larger infrastructure systems. - Highlights: • Novel algorithms developed for Bayesian network modeling of infrastructure systems. • Algorithm presented to compress information in conditional probability tables. • Updating algorithm presented to perform inference on compressed matrices. • Algorithms applied to example systems to investigate their performance. • Orders of magnitude savings in memory storage requirement demonstrated.

  14. Parallel algorithms for interactive manipulation of digital terrain models

    Science.gov (United States)

    Davis, E. W.; Mcallister, D. F.; Nagaraj, V.

    1988-01-01

    Interactive three-dimensional graphics applications, such as terrain data representation and manipulation, require extensive arithmetic processing. Massively parallel machines are attractive for this application since they offer high computational rates, and grid connected architectures provide a natural mapping for grid based terrain models. Presented here are algorithms for data movement on the massive parallel processor (MPP) in support of pan and zoom functions over large data grids. It is an extension of earlier work that demonstrated real-time performance of graphics functions on grids that were equal in size to the physical dimensions of the MPP. When the dimensions of a data grid exceed the processing array size, data is packed in the array memory. Windows of the total data grid are interactively selected for processing. Movement of packed data is needed to distribute items across the array for efficient parallel processing. Execution time for data movement was found to exceed that for arithmetic aspects of graphics functions. Performance figures are given for routines written in MPP Pascal.

  15. Parallelization of the model-based iterative reconstruction algorithm DIRA

    International Nuclear Information System (INIS)

    Oertenberg, A.; Sandborg, M.; Alm Carlsson, G.; Malusek, A.; Magnusson, M.

    2016-01-01

    New paradigms for parallel programming have been devised to simplify software development on multi-core processors and many-core graphical processing units (GPU). Despite their obvious benefits, the parallelization of existing computer programs is not an easy task. In this work, the use of the Open Multiprocessing (OpenMP) and Open Computing Language (OpenCL) frameworks is considered for the parallelization of the model-based iterative reconstruction algorithm DIRA with the aim to significantly shorten the code's execution time. Selected routines were parallelized using OpenMP and OpenCL libraries; some routines were converted from MATLAB to C and optimised. Parallelization of the code with the OpenMP was easy and resulted in an overall speedup of 15 on a 16-core computer. Parallelization with OpenCL was more difficult owing to differences between the central processing unit and GPU architectures. The resulting speedup was substantially lower than the theoretical peak performance of the GPU; the cause was explained. (authors)

  16. Solutions of matrix models in the DIII generator ensemble

    OpenAIRE

    Roussel, Harold

    1994-01-01

    In this paper we solve two matrix models, using standard and new techniques. The two models are represented by special form of antisymmetric matrices and are classified in the DIII generator ensemble. It is shown that, in the double scaling limit, their free energy has the same behavior as previous models describing oriented and unoriented surfaces. We also found an additional solution for the first model.

  17. Multi-objective optimization algorithms for mixed model assembly line balancing problem with parallel workstations

    Directory of Open Access Journals (Sweden)

    Masoud Rabbani

    2016-12-01

    Full Text Available This paper deals with mixed model assembly line (MMAL balancing problem of type-I. In MMALs several products are made on an assembly line while the similarity of these products is so high. As a result, it is possible to assemble several types of products simultaneously without any additional setup times. The problem has some particular features such as parallel workstations and precedence constraints in dynamic periods in which each period also effects on its next period. The research intends to reduce the number of workstations and maximize the workload smoothness between workstations. Dynamic periods are used to determine all variables in different periods to achieve efficient solutions. A non-dominated sorting genetic algorithm (NSGA-II and multi-objective particle swarm optimization (MOPSO are used to solve the problem. The proposed model is validated with GAMS software for small size problem and the performance of the foregoing algorithms is compared with each other based on some comparison metrics. The NSGA-II outperforms MOPSO with respect to some comparison metrics used in this paper, but in other metrics MOPSO is better than NSGA-II. Finally, conclusion and future research is provided.

  18. "Updates to Model Algorithms & Inputs for the Biogenic Emissions Inventory System (BEIS) Model"

    Science.gov (United States)

    We have developed new canopy emission algorithms and land use data for BEIS. Simulations with BEIS v3.4 and these updates in CMAQ v5.0.2 are compared these changes to the Model of Emissions of Gases and Aerosols from Nature (MEGAN) and evaluated the simulations against observatio...

  19. Rectangular Full Packed Format for Cholesky's Algorithm: Factorization, Solution, and Inversion

    DEFF Research Database (Denmark)

    Gustavson, Fred G.; Wasniewski, Jerzy; Dongarra, Jack J

    2010-01-01

    of the storage space but provide high performance via the use of Level 3 BLAS. Standard packed format arrays fully utilize storage (array space) but provide low performance as there is no Level 3 packed BLAS. We combine the good features of packed and full storage using RFPF to obtain high performance via using...... Level 3 BLAS as RFPF is a standard full-format representation. Also, RFPF requires exactly the same minimal storage as packed the format. Each LAPACK full and/or packed triangular, symmetric, and Hermitian routine becomes a single new RFPF routine based on eight possible data layouts of RFPF. This new...... RFPF routine usually consists of two calls to the corresponding LAPACK full-format routine and two calls to Level 3 BLAS routines. This means no new software is required. As examples, we present LAPACK routines for Cholesky factorization, Cholesky solution, and Cholesky inverse computation in RFPF...

  20. An asymptotic solution to a passive biped walker model

    Science.gov (United States)

    Yudaev, Sergey A.; Rachinskii, Dmitrii; Sobolev, Vladimir A.

    2017-02-01

    We consider a simple model of a passive dynamic biped robot walker with point feet and legs without knee. The model is a switched system, which includes an inverted double pendulum. Robot’s gait and its stability depend on parameters such as the slope of the ramp, the length of robot’s legs, and the mass distribution along the legs. We present an asymptotic solution of the model. The first correction to the zero order approximation is shown to agree with the numerical solution for a limited parameter range.

  1. Thermodynamic Models from Fluctuation Solution Theory Analysis of Molecular Simulations

    DEFF Research Database (Denmark)

    Christensen, Steen; Peters, Günther H.j.; Hansen, Flemming Yssing

    2007-01-01

    Fluctuation solution theory (FST) is employed to analyze results of molecular dynamics (MD) simulations of liquid mixtures. The objective is to generate parameters for macroscopic GE-models, here the modified Margules model. We present a strategy for choosing the number of parameters included...

  2. Modeling of CO2 absorber using an AMP solution

    DEFF Research Database (Denmark)

    Gabrielsen, Jostein; Michelsen, Michael Locht; Stenby, Erling Halfdan

    2006-01-01

    Abstract: An explicit model for carbon dioxide (CO2) solubility in an aqueous solution of 2-amino-2-methyl-1-propanol (AMP) has been proposed and an expression for the heat of absorption of CO2 has been developed as a function of loading and temperature. A rate-based steady-state model for CO2...

  3. A GROOVE Solution for the BPMN to BPEL Model Transformation

    NARCIS (Netherlands)

    de Mol, M.J.; Zimakova, M.V.

    In this paper we present a solution of a model transformation between two standard languages for business process modeling BPMN and BPEL, using the GROOVE tool set. GROOVE is a tool for graph transformations that uses directed, edge labelled simple graphs and the SPO approach [Ren04]. Given a graph

  4. Integrated algorithms for RFID-based multi-sensor indoor/outdoor positioning solutions

    Science.gov (United States)

    Zhu, Mi.; Retscher, G.; Zhang, K.

    2011-12-01

    Position information is very important as people need it almost everywhere all the time. However, it is a challenging task to provide precise positions indoor/outdoor seamlessly. Outdoor positioning has been widely studied and accurate positions can usually be achieved by well developed GPS techniques but these techniques are difficult to be used indoors since GPS signal reception is limited. The alternative techniques that can be used for indoor positioning include, to name a few, Wireless Local Area Network (WLAN), bluetooth and Ultra Wideband (UWB) etc.. However, all of these have limitations. The main objectives of this paper are to investigate and develop algorithms for a low-cost and portable indoor personal positioning system using Radio Frequency Identification (RFID) and its integration with other positioning systems. An RFID system consists of three components, namely a control unit, an interrogator and a transponder that transmits data and communicates with the reader. An RFID tag can be incorporated into a product, animal or person for the purpose of identification and tracking using radio waves. In general, for RFID positioning in urban and indoor environments three different methods can be used, including cellular positioning, trilateration and location fingerprinting. In addition, the integration of RFID with other technologies is also discussed in this paper. A typical combination is to integrate RFID with relative positioning technologies such as MEMS INS to bridge the gaps between RFID tags for continuous positioning applications. Experiments are shown to demonstrate the improvements of integrating multiple sensors with RFID which can be employed successfully for personal positioning.

  5. Algorithms for a parallel implementation of Hidden Markov Models with a small state space

    DEFF Research Database (Denmark)

    Nielsen, Jesper; Sand, Andreas

    2011-01-01

    Two of the most important algorithms for Hidden Markov Models are the forward and the Viterbi algorithms. We show how formulating these using linear algebra naturally lends itself to parallelization. Although the obtained algorithms are slow for Hidden Markov Models with large state spaces......, they require very little communication between processors, and are fast in practice on models with a small state space. We have tested our implementation against two other imple- mentations on artificial data and observe a speed-up of roughly a factor of 5 for the forward algorithm and more than 6...... for the Viterbi algorithm. We also tested our algorithm in the Coalescent Hidden Markov Model framework, where it gave a significant speed-up....

  6. Large time periodic solutions to coupled chemotaxis-fluid models

    Science.gov (United States)

    Jin, Chunhua

    2017-12-01

    In this paper, we deal with the time periodic problem to coupled chemotaxis-fluid models. We prove the existence of large time periodic strong solutions for the full chemotaxis-Navier-Stokes system in spatial dimension N=2, and the existence of large time periodic strong solutions for the chemotaxis-Stokes system in spatial dimension N=3. On the basis of these, the regularity of the solutions can be further improved. More precisely speaking, if the time periodic source g and the potential force \

  7. A Mesoscopic Model for Protein-Protein Interactions in Solution

    OpenAIRE

    Lund, Mikael; Jönsson, Bo

    2003-01-01

    Protein self-association may be detrimental in biological systems, but can be utilized in a controlled fashion for protein crystallization. It is hence of considerable interest to understand how factors like solution conditions prevent or promote aggregation. Here we present a computational model describing interactions between protein molecules in solution. The calculations are based on a molecular description capturing the detailed structure of the protein molecule using x-ray or nuclear ma...

  8. Jacobian elliptic wave solutions in an anharmonic molecular crystal model

    International Nuclear Information System (INIS)

    Teh, C.G.R.; Lee, B.S.; Koo, W.K.

    1997-07-01

    Explicit Jacobian elliptic wave solutions are found in the anharmonic molecular crystal model for both the continuum limit and discrete modes. This class of wave solutions include the famous pulse-like and kink-like solitary modes. We would also like to report on the existence of some highly discrete staggered solitary wave modes not found in the continuum limit. (author). 9 refs, 1 fig

  9. Models and Algorithms for Production Planning and Scheduling in Foundries – Current State and Development Perspectives

    Directory of Open Access Journals (Sweden)

    A. Stawowy

    2012-04-01

    Full Text Available Mathematical programming, constraint programming and computational intelligence techniques, presented in the literature in the field of operations research and production management, are generally inadequate for planning real-life production process. These methods are in fact dedicated to solving the standard problems such as shop floor scheduling or lot-sizing, or their simple combinations such as scheduling with batching. Whereas many real-world production planning problems require the simultaneous solution of several problems (in addition to task scheduling and lot-sizing, the problems such as cutting, workforce scheduling, packing and transport issues, including the problems that are difficult to structure. The article presents examples and classification of production planning and scheduling systems in the foundry industry described in the literature, and also outlines the possible development directions of models and algorithms used in such systems.

  10. A new parallel DNA algorithm to solve the task scheduling problem based on inspired computational model.

    Science.gov (United States)

    Wang, Zhaocai; Ji, Zuwen; Wang, Xiaoming; Wu, Tunhua; Huang, Wei

    2017-12-01

    As a promising approach to solve the computationally intractable problem, the method based on DNA computing is an emerging research area including mathematics, computer science and molecular biology. The task scheduling problem, as a well-known NP-complete problem, arranges n jobs to m individuals and finds the minimum execution time of last finished individual. In this paper, we use a biologically inspired computational model and describe a new parallel algorithm to solve the task scheduling problem by basic DNA molecular operations. In turn, we skillfully design flexible length DNA strands to represent elements of the allocation matrix, take appropriate biological experiment operations and get solutions of the task scheduling problem in proper length range with less than O(n 2 ) time complexity. Copyright © 2017. Published by Elsevier B.V.

  11. Leakage detection algorithm integrating water distribution networks hydraulic model

    CSIR Research Space (South Africa)

    Adedeji, K

    2017-06-01

    Full Text Available and estimation is vital for effective water service. For effective detection of background leakages, a hydraulic analysis of flow characteristics in water piping networks is indispensable for appraising such type of leakage. A leakage detection algorithm...

  12. COUPLING OF CORONAL AND HELIOSPHERIC MAGNETOHYDRODYNAMIC MODELS: SOLUTION COMPARISONS AND VERIFICATION

    Energy Technology Data Exchange (ETDEWEB)

    Merkin, V. G. [The Johns Hopkins University Applied Physics Laboratory, Laurel, MD 20723 (United States); Lionello, R.; Linker, J.; Török, T.; Downs, C. [Predictive Science, Inc., San Diego, CA 92121 (United States); Lyon, J. G., E-mail: slava.merkin@jhuapl.edu [Department of Physics and Astronomy, Dartmouth College, Hanover, NH 03755 (United States)

    2016-11-01

    Two well-established magnetohydrodynamic (MHD) codes are coupled to model the solar corona and the inner heliosphere. The corona is simulated using the MHD algorithm outside a sphere (MAS) model. The Lyon–Fedder–Mobarry (LFM) model is used in the heliosphere. The interface between the models is placed in a spherical shell above the critical point and allows both models to work in either a rotating or an inertial frame. Numerical tests are presented examining the coupled model solutions from 20 to 50 solar radii. The heliospheric simulations are run with both LFM and the MAS extension into the heliosphere, and use the same polytropic coronal MAS solutions as the inner boundary condition. The coronal simulations are performed for idealized magnetic configurations, with an out-of-equilibrium flux rope inserted into an axisymmetric background, with and without including the solar rotation. The temporal evolution at the inner boundary of the LFM and MAS solutions is shown to be nearly identical, as are the steady-state background solutions, prior to the insertion of the flux rope. However, after the coronal mass ejection has propagated through the significant portion of the simulation domain, the heliospheric solutions diverge. Additional simulations with different resolution are then performed and show that the MAS heliospheric solutions approach those of LFM when run with progressively higher resolution. Following these detailed tests, a more realistic simulation driven by the thermodynamic coronal MAS is presented, which includes solar rotation and an azimuthally asymmetric background and extends to the Earth’s orbit.

  13. Location-allocation models and new solution methodologies in telecommunication networks

    Science.gov (United States)

    Dinu, S.; Ciucur, V.

    2016-08-01

    When designing a telecommunications network topology, three types of interdependent decisions are combined: location, allocation and routing, which are expressed by the following design considerations: how many interconnection devices - consolidation points/concentrators should be used and where should they be located; how to allocate terminal nodes to concentrators; how should the voice, video or data traffic be routed and what transmission links (capacitated or not) should be built into the network. Including these three components of the decision into a single model generates a problem whose complexity makes it difficult to solve. A first method to address the overall problem is the sequential one, whereby the first step deals with the location-allocation problem and based on this solution the subsequent sub-problem (routing the network traffic) shall be solved. The issue of location and allocation in a telecommunications network, called "The capacitated concentrator location- allocation - CCLA problem" is based on one of the general location models on a network in which clients/demand nodes are the terminals and facilities are the concentrators. Like in a location model, each client node has a demand traffic, which must be served, and the facilities can serve these demands within their capacity limit. In this study, the CCLA problem is modeled as a single-source capacitated location-allocation model whose optimization objective is to determine the minimum network cost consisting of fixed costs for establishing the locations of concentrators, costs for operating concentrators and costs for allocating terminals to concentrators. The problem is known as a difficult combinatorial optimization problem for which powerful algorithms are required. Our approach proposes a Fuzzy Genetic Algorithm combined with a local search procedure to calculate the optimal values of the location and allocation variables. To confirm the efficiency of the proposed algorithm with respect

  14. SOLA-VOF: a solution algorithm for transient fluid flow with multiple free boundaries

    International Nuclear Information System (INIS)

    Nichols, B.D.; Hirt, C.W.; Hotchkiss, R.S.

    1980-08-01

    In this report a simple, but powerful, computer program is presented for the solution of two-dimensional transient fluid flow with free boundaries. The SOLA-VOF program, which is based on the concept of a fractional volume of fluid (VOF), is more flexible and efficient than other methods for treating arbitrary free boundaries. SOLA-VOF has a variety of user options that provide capabilities for a wide range of applications. Its basic mode of operation is for single fluid calculations having multiple free surfaces. However, SOLA-VOF can also be used for calculations involving two fluids separated by a sharp interface. In either case, the fluids may be treated as incompressible or as having limited compressibility. Surface tension forces with wall adhesion are permitted in both cases. Internal obstacles may be defined by blocking out any desired combination of cells in the mesh, which is composed of rectangular cells of variable size. SOLA-VOF is an easy-to-use program. Its logical parts are isolated in separate subroutines, and numerous special features have been included to simplify its operation, such as an automatic time-step control, a flexible mesh generator, extensive output capabilities, a variety of optional boundary conditions, and instructive internal documentation

  15. A molecular-thermodynamic model for polyelectrolyte solutions

    Energy Technology Data Exchange (ETDEWEB)

    Jiang, J.; Liu, H.; Hu, Y. [Thermodynamics Research Laboratory, East China University of Science and Technology, Shanghai 200237 (China); Prausnitz, J.M. [Department of Chemical Engineering, University of California, Berkeley, and Chemical Sciences Division, Lawrence Berkeley Laboratory, University of California, Berkeley, California 94720 (United States)

    1998-01-01

    Polyelectrolyte solutions are modeled as freely tangent-jointed, charged hard-sphere chains and corresponding counterions in a continuum medium with permitivity {var_epsilon}. By adopting the sticky-point model, the Helmholtz function for polyelectrolyte solutions is derived through the r-particle cavity-correlation function (CCF) for chains of sticky, charged hard spheres. The r-CCF is approximated by a product of effective nearest-neighbor two-particle CCFs; these are determined from the hypernetted-chain and mean-spherical closures (HNC/MSA) inside and outside the hard core, respectively, for the integral equation theory for electrolytes. The colligative properties are given as explicit functions of a scaling parameter {Gamma} that can be estimated by a simple iteration procedure. Osmotic pressures, osmotic coefficients, and activity coefficients are calculated for model solutions with various chain lengths. They are in good agreement with molecular simulation and experimental results. {copyright} {ital 1998 American Institute of Physics.}

  16. A biogeography-based optimization algorithm with mutation strategies for model parameter estimation of solar and fuel cells

    International Nuclear Information System (INIS)

    Niu, Qun; Zhang, Letian; Li, Kang

    2014-01-01

    Highlights: • Solar cell and PEM fuel cell parameter estimations are investigated in the paper. • A new biogeography-based method (BBO-M) is proposed for cell parameter estimations. • In BBO-M, two mutation operators are designed to enhance optimization performance. • BBO-M provides a competitive alternative in cell parameter estimation problems. - Abstract: Mathematical models are useful tools for simulation, evaluation, optimal operation and control of solar cells and proton exchange membrane fuel cells (PEMFCs). To identify the model parameters of these two type of cells efficiently, a biogeography-based optimization algorithm with mutation strategies (BBO-M) is proposed. The BBO-M uses the structure of biogeography-based optimization algorithm (BBO), and both the mutation motivated from the differential evolution (DE) algorithm and the chaos theory are incorporated into the BBO structure for improving the global searching capability of the algorithm. Numerical experiments have been conducted on ten benchmark functions with 50 dimensions, and the results show that BBO-M can produce solutions of high quality and has fast convergence rate. Then, the proposed BBO-M is applied to the model parameter estimation of the two type of cells. The experimental results clearly demonstrate the power of the proposed BBO-M in estimating model parameters of both solar and fuel cells

  17. Research on compressive sensing reconstruction algorithm based on total variation model

    Science.gov (United States)

    Gao, Yu-xuan; Sun, Huayan; Zhang, Tinghua; Du, Lin

    2017-12-01

    Compressed sensing for breakthrough Nyquist sampling theorem provides a strong theoretical , making compressive sampling for image signals be carried out simultaneously. In traditional imaging procedures using compressed sensing theory, not only can it reduces the storage space, but also can reduce the demand for detector resolution greatly. Using the sparsity of image signal, by solving the mathematical model of inverse reconfiguration, realize the super-resolution imaging. Reconstruction algorithm is the most critical part of compression perception, to a large extent determine the accuracy of the reconstruction of the image.The reconstruction algorithm based on the total variation (TV) model is more suitable for the compression reconstruction of the two-dimensional image, and the better edge information can be obtained. In order to verify the performance of the algorithm, Simulation Analysis the reconstruction result in different coding mode of the reconstruction algorithm based on the TV reconstruction algorithm. The reconstruction effect of the reconfigurable algorithm based on TV based on the different coding methods is analyzed to verify the stability of the algorithm. This paper compares and analyzes the typical reconstruction algorithm in the same coding mode. On the basis of the minimum total variation algorithm, the Augmented Lagrangian function term is added and the optimal value is solved by the alternating direction method.Experimental results show that the reconstruction algorithm is compared with the traditional classical algorithm based on TV has great advantages, under the low measurement rate can be quickly and accurately recovers target image.

  18. Global solution for a chemotactic haptotactic model of cancer invasion

    Science.gov (United States)

    Tao, Youshan; Wang, Mingjun

    2008-10-01

    This paper deals with a mathematical model of cancer invasion of tissue recently proposed by Chaplain and Lolas. The model consists of a reaction-diffusion-taxis partial differential equation (PDE) describing the evolution of tumour cell density, a reaction-diffusion PDE governing the evolution of the proteolytic enzyme concentration and an ordinary differential equation modelling the proteolysis of the extracellular matrix (ECM). In addition to random motion, the tumour cells are directed not only by haptotaxis (cellular locomotion directed in response to a concentration gradient of adhesive molecules along the ECM) but also by chemotaxis (cellular locomotion directed in response to a concentration gradient of the diffusible proteolytic enzyme). In one space dimension, the global existence and uniqueness of a classical solution to this combined chemotactic-haptotactic model is proved for any chemotactic coefficient χ > 0. In two and three space dimensions, the global existence is proved for small χ/μ (where μ is the logistic growth rate of the tumour cells). The fundamental point of proof is to raise the regularity of a solution from L1 to Lp (p > 1). Furthermore, the existence of blow-up solutions to a sub-model in two space dimensions for large χ shows, to some extent, that the condition that χ/μ is small is necessary for the global existence of a solution to the full model.

  19. Modeling and inversion Matlab algorithms for resistivity, induced polarization and seismic data

    Science.gov (United States)

    Karaoulis, M.; Revil, A.; Minsley, B. J.; Werkema, D. D.

    2011-12-01

    M. Karaoulis (1), D.D. Werkema (3), A. Revil (1,2), A., B. Minsley (4), (1) Colorado School of Mines, Dept. of Geophysics, Golden, CO, USA. (2) ISTerre, CNRS, UMR 5559, Université de Savoie, Equipe Volcan, Le Bourget du Lac, France. (3) U.S. EPA, ORD, NERL, ESD, CMB, Las Vegas, Nevada, USA . (4) USGS, Federal Center, Lakewood, 10, 80225-0046, CO. Abstract We propose 2D and 3D forward modeling and inversion package for DC resistivity, time domain induced polarization (IP), frequency-domain IP, and seismic refraction data. For the resistivity and IP case, discretization is based on rectangular cells, where each cell has as unknown resistivity in the case of DC modelling, resistivity and chargeability in the time domain IP modelling, and complex resistivity in the spectral IP modelling. The governing partial-differential equations are solved with the finite element method, which can be applied to both real and complex variables that are solved for. For the seismic case, forward modeling is based on solving the eikonal equation using a second-order fast marching method. The wavepaths are materialized by Fresnel volumes rather than by conventional rays. This approach accounts for complicated velocity models and is advantageous because it considers frequency effects on the velocity resolution. The inversion can accommodate data at a single time step, or as a time-lapse dataset if the geophysical data are gathered for monitoring purposes. The aim of time-lapse inversion is to find the change in the velocities or resistivities of each model cell as a function of time. Different time-lapse algorithms can be applied such as independent inversion, difference inversion, 4D inversion, and 4D active time constraint inversion. The forward algorithms are benchmarked against analytical solutions and inversion results are compared with existing ones. The algorithms are packaged as Matlab codes with a simple Graphical User Interface. Although the code is parallelized for multi

  20. Making the error-controlling algorithm of observable operator models constructive.

    Science.gov (United States)

    Zhao, Ming-Jie; Jaeger, Herbert; Thon, Michael

    2009-12-01

    Observable operator models (OOMs) are a class of models for stochastic processes that properly subsumes the class that can be modeled by finite-dimensional hidden Markov models (HMMs). One of the main advantages of OOMs over HMMs is that they admit asymptotically correct learning algorithms. A series of learning algorithms has been developed, with increasing computational and statistical efficiency, whose recent culmination was the error-controlling (EC) algorithm developed by the first author. The EC algorithm is an iterative, asymptotically correct algorithm that yields (and minimizes) an assured upper bound on the modeling error. The run time is faster by at least one order of magnitude than EM-based HMM learning algorithms and yields significantly more accurate models than the latter. Here we present a significant improvement of the EC algorithm: the constructive error-controlling (CEC) algorithm. CEC inherits from EC the main idea of minimizing an upper bound on the modeling error but is constructive where EC needs iterations. As a consequence, we obtain further gains in learning speed without loss in modeling accuracy.

  1. Parallel Algorithm for Solving TOV Equations for Sequence of Cold and Dense Nuclear Matter Models

    Science.gov (United States)

    Ayriyan, Alexander; Buša, Ján; Grigorian, Hovik; Poghosyan, Gevorg

    2018-04-01

    We have introduced parallel algorithm simulation of neutron star configurations for set of equation of state models. The performance of the parallel algorithm has been investigated for testing set of EoS models on two computational systems. It scales when using with MPI on modern CPUs and this investigation allowed us also to compare two different types of computational nodes.

  2. A genetic algorithm for a bi-objective mathematical model for dynamic virtual cell formation problem

    Science.gov (United States)

    Moradgholi, Mostafa; Paydar, Mohammad Mahdi; Mahdavi, Iraj; Jouzdani, Javid

    2016-05-01

    Nowadays, with the increasing pressure of the competitive business environment and demand for diverse products, manufacturers are force to seek for solutions that reduce production costs and rise product quality. Cellular manufacturing system (CMS), as a means to this end, has been a point of attraction to both researchers and practitioners. Limitations of cell formation problem (CFP), as one of important topics in CMS, have led to the introduction of virtual CMS (VCMS). This research addresses a bi-objective dynamic virtual cell formation problem (DVCFP) with the objective of finding the optimal formation of cells, considering the material handling costs, fixed machine installation costs and variable production costs of machines and workforce. Furthermore, we consider different skills on different machines in workforce assignment in a multi-period planning horizon. The bi-objective model is transformed to a single-objective fuzzy goal programming model and to show its performance; numerical examples are solved using the LINGO software. In addition, genetic algorithm (GA) is customized to tackle large-scale instances of the problems to show the performance of the solution method.

  3. Using Hadoop MapReduce for Parallel Genetic Algorithms: A Comparison of the Global, Grid and Island Models.

    Science.gov (United States)

    Ferrucci, Filomena; Salza, Pasquale; Sarro, Federica

    2017-06-29

    The need to improve the scalability of Genetic Algorithms (GAs) has motivated the research on Parallel Genetic Algorithms (PGAs), and different technologies and approaches have been used. Hadoop MapReduce represents one of the most mature technologies to develop parallel algorithms. Based on the fact that parallel algorithms introduce communication overhead, the aim of the present work is to understand if, and possibly when, the parallel GAs solutions using Hadoop MapReduce show better performance than sequential versions in terms of execution time. Moreover, we are interested in understanding which PGA model can be most effective among the global, grid, and island models. We empirically assessed the performance of these three parallel models with respect to a sequential GA on a software engineering problem, evaluating the execution time and the achieved speedup. We also analysed the behaviour of the parallel models in relation to the overhead produced by the use of Hadoop MapReduce and the GAs' computational effort, which gives a more machine-independent measure of these algorithms. We exploited three problem instances to differentiate the computation load and three cluster configurations based on 2, 4, and 8 parallel nodes. Moreover, we estimated the costs of the execution of the experimentation on a potential cloud infrastructure, based on the pricing of the major commercial cloud providers. The empirical study revealed that the use of PGA based on the island model outperforms the other parallel models and the sequential GA for all the considered instances and clusters. Using 2, 4, and 8 nodes, the island model achieves an average speedup over the three datasets of 1.8, 3.4, and 7.0 times, respectively. Hadoop MapReduce has a set of different constraints that need to be considered during the design and the implementation of parallel algorithms. The overhead of data store (i.e., HDFS) accesses, communication, and latency requires solutions that reduce data store

  4. Phase-field model and its numerical solution for coring and microstructure evolution studies in alloys

    Science.gov (United States)

    Turchi, Patrice E. A.; Fattebert, Jean-Luc; Dorr, Milo R.; Wickett, Michael E.; Belak, James F.

    2011-03-01

    We describe an algorithm for the numerical solution of a phase-field model (PFM) of microstructure evolution in alloys using physical parameters from thermodynamic (CALPHAD) and kinetic databases. The coupled system of PFM equations includes a local order parameter, a quaternion representation of local crystal orientation and a species composition parameter. Time evolution of microstructures and alloy composition is obtained using an implicit time integration of the system. Physical parameters in databases can be obtained either through experiment or first-principles calculations. Application to coring studies and microstructure evolution of Au-Ni will be presented. Prepared by LLNL under Contract DE-AC52-07NA27344

  5. Evaluation of unsaturated-zone solute-transport models for studies of agricultural chemicals

    Science.gov (United States)

    Nolan, Bernard T.; Bayless, E. Randall; Green, Christopher T.; Garg, Sheena; Voss, Frank D.; Lampe, David C.; Barbash, Jack E.; Capel, Paul D.; Bekins, Barbara A.

    2005-01-01

    Seven unsaturated-zone solute-transport models were tested with two data sets to select models for use by the Agricultural Chemical Team of the U.S. Geological Survey's National Water-Quality Assessment Program. The data sets were from a bromide tracer test near Merced, California, and an atrazine study in the White River Basin, Indiana. In this study the models are designated either as complex or simple based on the water flux algorithm. The complex models, HYDRUS2D, LEACHP, RZWQM, and VS2DT, use Richards' equation to simulate water flux and are well suited to process understanding. The simple models, CALF, GLEAMS, and PRZM, use a tipping-bucket algorithm and are more amenable to extrapolation because they require fewer input parameters. The purpose of this report is not to endorse a particular model, but to describe useful features, potential capabilities, and possible limitations that emerged from working with the model input data sets. More rigorous assessment of model applicability involves proper calibration, which was beyond the scope of this study.

  6. Dual-Model Reverse CKF Algorithm in Cooperative Navigation for USV

    Directory of Open Access Journals (Sweden)

    Bo Xu

    2014-01-01

    Full Text Available As one of the most promising research directions, cooperative location with high precision and low-cost IMU is becoming an emerging research topic in many positioning fields. Low-cost MEMS/DVL is a preferred solution for dead-reckoning in multi-USV cooperative network. However, large misalignment angles and large gyro drift coexist in low-cost MEMS that leads to the poor observability. Based on cubature Kalman filter (CKF algorithm that has access to high accuracy and relative small computation, dual-model filtering scheme is proposed. It divides the whole process into two subsections that cut off the coupling relations and improve the observability of MEMS errors: it first estimates large misalignment angle and then estimates the gyro drift. Furthermore, to improve the convergence speed of large misalignment angle estimated in the first subsection, “time reversion” concept is introduced. It uses a short period time to forward and backward several times to improve convergence speed effectively. Finally, simulation analysis and experimental verification is conducted. Simulation and experimental results show that the algorithm can effectively improve the cooperative navigation performance.

  7. Mathematical model and coordination algorithms for ensuring complex security of an organization

    Science.gov (United States)

    Novoseltsev, V. I.; Orlova, D. E.; Dubrovin, A. S.; Irkhin, V. P.

    2018-03-01

    The mathematical model of coordination when ensuring complex security of the organization is considered. On the basis of use of a method of casual search three types of algorithms of effective coordination adequate to mismatch level concerning security are developed: a coordination algorithm at domination of instructions of the coordinator; a coordination algorithm at domination of decisions of performers; a coordination algorithm at parity of interests of the coordinator and performers. Assessment of convergence of the algorithms considered above it was made by carrying out a computing experiment. The described algorithms of coordination have property of convergence in the sense stated above. And, the following regularity is revealed: than more simply in the structural relation the algorithm, for the smaller number of iterations is provided to those its convergence.

  8. A Path Planning Algorithm using Generalized Potential Model for Hyper- Redundant Robots with 2-DOF Joints

    Directory of Open Access Journals (Sweden)

    Chien-Chou Lin

    2011-06-01

    Full Text Available This paper proposes a potential‐based path planning algorithm of articulated robots with 2‐DOF joints. The algorithm is an extension of a previous algorithm developed for 3‐DOF joints. While 3‐DOF joints result in a more straightforward potential minimization algorithm, 2‐DOF joints are obviously more practical for active operations. The proposed approach computes repulsive force and torque between charged objects by using generalized potential model. A collision‐free path can be obtained by locally adjusting the robot configuration to search for minimum potential configurations using these force and torque. The optimization of path safeness, through the innovative potential minimization algorithm, makes the proposed approach unique. In order to speedup the computation, a sequential planning strategy is adopted. Simulation results show that the proposed algorithm works well compared with 3‐DOF‐joint algorithm, in terms of collision avoidance and computation efficiency.

  9. A Decomposition Algorithm for Mean-Variance Economic Model Predictive Control of Stochastic Linear Systems

    DEFF Research Database (Denmark)

    Sokoler, Leo Emil; Dammann, Bernd; Madsen, Henrik

    2014-01-01

    This paper presents a decomposition algorithm for solving the optimal control problem (OCP) that arises in Mean-Variance Economic Model Predictive Control of stochastic linear systems. The algorithm applies the alternating direction method of multipliers to a reformulation of the OCP that decompo......This paper presents a decomposition algorithm for solving the optimal control problem (OCP) that arises in Mean-Variance Economic Model Predictive Control of stochastic linear systems. The algorithm applies the alternating direction method of multipliers to a reformulation of the OCP...

  10. Inverse modelling in estimating soil hydraulic functions: a Genetic Algorithm approach

    Science.gov (United States)

    Ines, A. V. M.; Droogers, P.

    The practical application of simulation models in the field is sometimes hindered by the difficulty of deriving the soil hydraulic properties of the study area. The procedure so-called inverse modelling has been investigated in many studies to address the problem where most of the studies were limited to hypothetical soil profile and soil core samples in the laboratory. Often, the numerical approach called forward-backward simulation is employed to generate synthetic data then added with random errors to mimic the real-world condition. Inverse modelling is used to backtrack the expected values of the parameters. This study explored the potential of a Genetic Algorithm (GA) to estimate inversely the soil hydraulic functions in the unsaturated zone. Lysimeter data from a wheat experiment in India were used in the analysis. Two cases were considered: (1) a numerical case where the forward-backward approach was employed and (2) the experimental case where the real data from the lysimeter experiment were used. Concurrently, the use of soil water, evapotranspiration (ET) and the combination of both were investigated as criteria in the inverse modelling. Results showed that using soil water as a criterion provides more accurate parameter estimates than using ET. However, from a practical point of view, ET is more attractive as it can be obtained with reasonable accuracy on a regional scale from remote sensing observations. The experimental study proved that the forward-backward approach does not take into account the effects of model errors. The formulation of the problem is found to be critical for a successful parameter estimation. The sensitivity of parameters to the objective function and their zone of influence in the soil column are major determinants in the solution. Generally, their effects sometimes lead to non-uniqueness in the solution but to some extent are partly handled by GA. Overall, it was concluded that the GA approach is promising to the inverse problem

  11. Inverse modelling in estimating soil hydraulic functions: a Genetic Algorithm approach

    Directory of Open Access Journals (Sweden)

    A. V. M. Ines

    2002-01-01

    Full Text Available The practical application of simulation models in the field is sometimes hindered by the difficulty of deriving the soil hydraulic properties of the study area. The procedure so-called inverse modelling has been investigated in many studies to address the problem where most of the studies were limited to hypothetical soil profile and soil core samples in the laboratory. Often, the numerical approach called forward-backward simulation is employed to generate synthetic data then added with random errors to mimic the real-world condition. Inverse modelling is used to backtrack the expected values of the parameters. This study explored the potential of a Genetic Algorithm (GA to estimate inversely the soil hydraulic functions in the unsaturated zone. Lysimeter data from a wheat experiment in India were used in the analysis. Two cases were considered: (1 a numerical case where the forward-backward approach was employed and (2 the experimental case where the real data from the lysimeter experiment were used. Concurrently, the use of soil water, evapotranspiration (ET and the combination of both were investigated as criteria in the inverse modelling. Results showed that using soil water as a criterion provides more accurate parameter estimates than using ET. However, from a practical point of view, ET is more attractive as it can be obtained with reasonable accuracy on a regional scale from remote sensing observations. The experimental study proved that the forward-backward approach does not take into account the effects of model errors. The formulation of the problem is found to be critical for a successful parameter estimation. The sensitivity of parameters to the objective function and their zone of influence in the soil column are major determinants in the solution. Generally, their effects sometimes lead to non-uniqueness in the solution but to some extent are partly handled by GA. Overall, it was concluded that the GA approach is promising to the

  12. A Multi-Scale Method for Dynamics Simulation in Continuum Solvent Models I: Finite-Difference Algorithm for Navier-Stokes Equation.

    Science.gov (United States)

    Xiao, Li; Cai, Qin; Li, Zhilin; Zhao, Hongkai; Luo, Ray

    2014-11-25

    A multi-scale framework is proposed for more realistic molecular dynamics simulations in continuum solvent models by coupling a molecular mechanics treatment of solute with a fluid mechanics treatment of solvent. This article reports our initial efforts to formulate the physical concepts necessary for coupling the two mechanics and develop a 3D numerical algorithm to simulate the solvent fluid via the Navier-Stokes equation. The numerical algorithm was validated with multiple test cases. The validation shows that the algorithm is effective and stable, with observed accuracy consistent with our design.

  13. Artificial neural network-genetic algorithm based optimization for the adsorption of methylene blue and brilliant green from aqueous solution by graphite oxide nanoparticle.

    Science.gov (United States)

    Ghaedi, M; Zeinali, N; Ghaedi, A M; Teimuori, M; Tashkhourian, J

    2014-05-05

    In this study, graphite oxide (GO) nano according to Hummers method was synthesized and subsequently was used for the removal of methylene blue (MB) and brilliant green (BG). The detail information about the structure and physicochemical properties of GO are investigated by different techniques such as XRD and FTIR analysis. The influence of solution pH, initial dye concentration, contact time and adsorbent dosage was examined in batch mode and optimum conditions was set as pH=7.0, 2 mg of GO and 10 min contact time. Employment of equilibrium isotherm models for description of adsorption capacities of GO explore the good efficiency of Langmuir model for the best presentation of experimental data with maximum adsorption capacity of 476.19 and 416.67 for MB and BG dyes in single solution. The analysis of adsorption rate at various stirring times shows that both dyes adsorption followed a pseudo second-order kinetic model with cooperation with interparticle diffusion model. Subsequently, the adsorption data as new combination of artificial neural network was modeled to evaluate and obtain the real conditions for fast and efficient removal of dyes. A three-layer artificial neural network (ANN) model is applicable for accurate prediction of dyes removal percentage from aqueous solution by GO following conduction of 336 experimental data. The network was trained using the obtained experimental data at optimum pH with different GO amount (0.002-0.008 g) and 5-40 mg/L of both dyes over contact time of 0.5-30 min. The ANN model was able to predict the removal efficiency with Levenberg-Marquardt algorithm (LMA), a linear transfer function (purelin) at output layer and a tangent sigmoid transfer function (tansig) at hidden layer with 10 and 11 neurons for MB and BG dyes, respectively. The minimum mean squared error (MSE) of 0.0012 and coefficient of determination (R(2)) of 0.982 were found for prediction and modeling of MB removal, while the respective value for BG was the

  14. Algorithms and Methods for High-Performance Model Predictive Control

    DEFF Research Database (Denmark)

    Frison, Gianluca

    routines employed in the numerical tests. The main focus of this thesis is on linear MPC problems. In this thesis, both the algorithms and their implementation are equally important. About the implementation, a novel implementation strategy for the dense linear algebra routines in embedded optimization...... is proposed, aiming at improving the computational performance in case of small matrices. About the algorithms, they are built on top of the proposed linear algebra, and they are tailored to exploit the high-level structure of the MPC problems, with special care on reducing the computational complexity....

  15. A hybrid algorithm and its applications to fuzzy logic modeling of nonlinear systems

    Science.gov (United States)

    Wang, Zhongjun

    System models allow us to simulate and analyze system dynamics efficiently. Most importantly, system models allow us to make prediction about system behaviors and to perform system parametric variation analysis without having to build the actual systems. The fuzzy logic modeling technique has been successfully applied in complex nonlinear system modeling such as unsteady aerodynamics modeling etc. recently. However, the current forward search algorithm to identify fuzzy logic model structures is very time-consuming. It is not unusual to spend several days or even a few weeks in computer CPU time to obtain better nonlinear system model structures by this forward search. Moreover, how to speed up the fuzzy logic model parameter identification process is also challenging when the number of influencing variables of nonlinear systems is large. To solve these problems, a hybrid algorithm for the nonlinear system modeling is proposed, formalized, implemented, and evaluated in this dissertation. By combining the fuzzy logic modeling technique with genetic algorithms, the developed hybrid algorithm is applied to both fuzzy logic model structure identification and model parameter identification. In the model structure identification process, the hybrid algorithm has the ability to find feasible structures more efficiently and effectively than the forward search. In the model parameter identification process (by using Newton gradient descent algorithm), the proposed hybrid algorithm incorporates genetic search algorithm to dynamically select convergence factors. It has the advantages of quick search yet maintains the monotonically convergent properties of the Newton gradient descent algorithm. To evaluate the properties of the developed hybrid algorithm, a nonlinear, unsteady aerodynamic normal force model with a complex system involving fourteen influencing variables is established from flight data. The results show that this hybrid algorithm can identify the aerodynamic

  16. Stability of subsystem solutions in agent-based models

    Science.gov (United States)

    Perc, Matjaž

    2018-01-01

    The fact that relatively simple entities, such as particles or neurons, or even ants or bees or humans, give rise to fascinatingly complex behaviour when interacting in large numbers is the hallmark of complex systems science. Agent-based models are frequently employed for modelling and obtaining a predictive understanding of complex systems. Since the sheer number of equations that describe the behaviour of an entire agent-based model often makes it impossible to solve such models exactly, Monte Carlo simulation methods must be used for the analysis. However, unlike pairwise interactions among particles that typically govern solid-state physics systems, interactions among agents that describe systems in biology, sociology or the humanities often involve group interactions, and they also involve a larger number of possible states even for the most simplified description of reality. This begets the question: when can we be certain that an observed simulation outcome of an agent-based model is actually stable and valid in the large system-size limit? The latter is key for the correct determination of phase transitions between different stable solutions, and for the understanding of the underlying microscopic processes that led to these phase transitions. We show that a satisfactory answer can only be obtained by means of a complete stability analysis of subsystem solutions. A subsystem solution can be formed by any subset of all possible agent states. The winner between two subsystem solutions can be determined by the average moving direction of the invasion front that separates them, yet it is crucial that the competing subsystem solutions are characterised by a proper composition and spatiotemporal structure before the competition starts. We use the spatial public goods game with diverse tolerance as an example, but the approach has relevance for a wide variety of agent-based models.

  17. ADVAN-style analytical solutions for common pharmacokinetic models.

    Science.gov (United States)

    Abuhelwa, Ahmad Y; Foster, David J R; Upton, Richard N

    2015-01-01

    The analytical solutions to compartmental pharmacokinetic models are well known, but have not been presented in a form that easily allows for complex dosing regimen and changes in covariate/parameter values that may occur at discrete times within and/or between dosing intervals. Laplace transforms were used to derive ADVAN-style analytical solutions for 1, 2, and 3 compartment pharmacokinetic linear models of intravenous and first-order absorption drug administration. The equations calculate the change in drug amounts in each compartment of the model over a time interval (t; t = t2 - t1) accounting for any dose or covariate events acting in the time interval. The equations were coded in the R language and used to simulate the time-course of drug amounts in each compartment of the systems. The equations were validated against commercial software [NONMEM (Beal, Sheiner, Boeckmann, & Bauer, 2009)] output to assess their capability to handle both complex dosage regimens and the effect of changes in covariate/parameter values that may occur at discrete times within or between dosing intervals. For all tested pharmacokinetic models, the time-course of drug amounts using the ADVAN-style analytical solutions were identical to NONMEM outputs to at least four significant figures, confirming the validity of the presented equations. To our knowledge, this paper presents the ADVAN-style equations for common pharmacokinetic models in the literature for the first time. The presented ADVAN-style equations overcome obstacles to implementing the classical analytical solutions in software, and have speed advantages over solutions using differential equation solvers. The equations presented in this paper fill a gap in the pharmacokinetic literature, and it is expected that these equations will facilitate the investigation of useful open-source software for modelling pharmacokinetic data. Copyright © 2015 Elsevier Inc. All rights reserved.

  18. A Decomposition Model for HPLC-DAD Data Set and Its Solution by Particle Swarm Optimization

    Directory of Open Access Journals (Sweden)

    Lizhi Cui

    2014-01-01

    Full Text Available This paper proposes a separation method, based on the model of Generalized Reference Curve Measurement and the algorithm of Particle Swarm Optimization (GRCM-PSO, for the High Performance Liquid Chromatography with Diode Array Detection (HPLC-DAD data set. Firstly, initial parameters are generated to construct reference curves for the chromatogram peaks of the compounds based on its physical principle. Then, a General Reference Curve Measurement (GRCM model is designed to transform these parameters to scalar values, which indicate the fitness for all parameters. Thirdly, rough solutions are found by searching individual target for every parameter, and reinitialization only around these rough solutions is executed. Then, the Particle Swarm Optimization (PSO algorithm is adopted to obtain the optimal parameters by minimizing the fitness of these new parameters given by the GRCM model. Finally, spectra for the compounds are estimated based on the optimal parameters and the HPLC-DAD data set. Through simulations and experiments, following conclusions are drawn: (1 the GRCM-PSO method can separate the chromatogram peaks and spectra from the HPLC-DAD data set without knowing the number of the compounds in advance even when severe overlap and white noise exist; (2 the GRCM-PSO method is able to handle the real HPLC-DAD data set.

  19. Model predictive control algorithms and their application to a continuous fermenter

    Directory of Open Access Journals (Sweden)

    R. G. SILVA

    1999-06-01

    Full Text Available In many continuous fermentation processes, the control objective is to maximize productivity per unit time. The optimum operational point in the steady state can be obtained by maximizing the productivity rate using feed substrate concentration as the independent variable with the equations of the static model as constraints. In the present study, three model-based control schemes have been developed and implemented for a continuous fermenter. The first method modifies the well-known dynamic matrix control (DMC algorithm by making it adaptive. The other two use nonlinear model predictive control algorithms (NMPC, nonlinear model predictive control for calculation of control actions. The NMPC1 algorithm, which uses orthogonal collocation in finite elements, acted similar to NMPC2, which uses equidistant collocation. These algorithms are compared with DMC. The results obtained show the good performance of nonlinear algorithms.

  20. Development and performance analysis of model-based fault detection and diagnosis algorithm

    International Nuclear Information System (INIS)

    Kim, Jung Taek; Park, Jae Chang; Lee, Jung Woon; Kim, Kyung Youn; Lee, In Soo; Kim, Bong Seok; Kang, Sook In

    2002-05-01

    It is important to note that an effective means to assure the reliability and security for the nuclear power plant is to detect and diagnose the faults (failures) as soon and as accurately as possible. The objective of the project is to develop model-based fault detection and diagnosis algorithm for the pressurized water reactor and evaluate the performance of the developed algorithm. The scope of the work can be classified into two categories. The one is state-space model-based FDD algorithm based on the interacting multiple model (IMM) algorithm. The other is input-output model-based FDD algorithm based on the ART neural network. Extensive computer simulations are carried out to evaluate the performance in terms of speed and accuracy

  1. Pole solution in six dimensions as a dimensional reduction model

    Science.gov (United States)

    Ichinose, Shoichi

    2002-01-01

    A solution with the pole configuration in six dimensions is analyzed. It is a dimensional reduction model of Randall-Sundrum type. The soliton configuration is induced by the bulk Higgs mechanism. The boundary condition is systematically solved up to the 6th order. The Riemann curvature is finite everywhere.

  2. LED-based Photometric Stereo: Modeling, Calibration and Numerical Solutions

    DEFF Research Database (Denmark)

    Quéau, Yvain; Durix, Bastien; Wu, Tao

    2018-01-01

    We conduct a thorough study of photometric stereo under nearby point light source illumination, from modeling to numerical solution, through calibration. In the classical formulation of photometric stereo, the luminous fluxes are assumed to be directional, which is very difficult to achieve in pr...

  3. Approximate Solutions of Interactive Dynamic Influence Diagrams Using Model Clustering

    DEFF Research Database (Denmark)

    Zeng, Yifeng; Doshi, Prashant; Qiongyu, Cheng

    2007-01-01

    Interactive dynamic influence diagrams (I-DIDs) offer a transparent and semantically clear representation for the sequential decision-making problem over multiple time steps in the presence of other interacting agents. Solving I-DIDs exactly involves knowing the solutions of possible models...

  4. Stationary solutions of multicomponent chiral and gauge models

    International Nuclear Information System (INIS)

    Chudnovsky, D.V.; Chudnovsky, G.V.

    1979-01-01

    The authors examine stationary solutions of completely integrable systems in (x, t) dimensions having infinitely many components. Among the cases under investigation are: (1) the infinite-component non-linear Schroedinger equation; (2) infinite component CPsup(Ω) or SU(N) sigma-models; (3) general gauge and chiral completely integrable systems. (Auth.)

  5. Interpolation solution of the single-impurity Anderson model

    International Nuclear Information System (INIS)

    Kuzemsky, A.L.

    1990-10-01

    The dynamical properties of the single-impurity Anderson model (SIAM) is studied using a novel Irreducible Green's Function method (IGF). The new solution for one-particle GF interpolating between the strong and weak correlation limits is obtained. The unified concept of relevant mean-field renormalizations is indispensable for strong correlation limit. (author). 21 refs

  6. Analysis and modeling of alkali halide aqueous solutions

    DEFF Research Database (Denmark)

    Kim, Sun Hyung; Anantpinijwatna, Amata; Kang, Jeong Won

    2016-01-01

    on calculations for various electrolyte properties of alkali halide aqueous solutions such as mean ionic activity coefficients, osmotic coefficients, and salt solubilities. The model covers highly nonideal electrolyte systems such as lithium chloride, lithium bromide and lithium iodide, that is, systems...

  7. Foam for Enhanced Oil Recovery : Modeling and Analytical Solutions

    NARCIS (Netherlands)

    Ashoori, E.

    2012-01-01

    Foam increases sweep in miscible- and immiscible-gas enhanced oil recovery by decreasing the mobility of gas enormously. This thesis is concerned with the simulations and analytical solutions for foam flow for the purpose of modeling foam EOR in a reservoir. For the ultimate goal of upscaling our

  8. Small-scale engagement model with arrivals: analytical solutions

    International Nuclear Information System (INIS)

    Engi, D.

    1977-04-01

    This report presents an analytical model of small-scale battles. The specific impetus for this effort was provided by a need to characterize hypothetical battles between guards at a nuclear facility and their potential adversaries. The solution procedure can be used to find measures of a number of critical parameters; for example, the win probabilities and the expected duration of the battle. Numerical solutions are obtainable if the total number of individual combatants on the opposing sides is less than 10. For smaller force size battles, with one or two combatants on each side, symbolic solutions can be found. The symbolic solutions express the output parameters abstractly in terms of symbolic representations of the input parameters while the numerical solutions are expressed as numerical values. The input parameters are derived from the probability distributions of the attrition and arrival processes. The solution procedure reduces to solving sets of linear equations that have been constructed from the input parameters. The approach presented in this report does not address the problems associated with measuring the inputs. Rather, this report attempts to establish a relatively simple structure within which small-scale battles can be studied

  9. Pollutant source identification model for water pollution incidents in small straight rivers based on genetic algorithm

    Science.gov (United States)

    Zhang, Shou-ping; Xin, Xiao-kang

    2017-07-01

    Identification of pollutant sources for river pollution incidents is an important and difficult task in the emergency rescue, and an intelligent optimization method can effectively compensate for the weakness of traditional methods. An intelligent model for pollutant source identification has been established using the basic genetic algorithm (BGA) as an optimization search tool and applying an analytic solution formula of one-dimensional unsteady water quality equation to construct the objective function. Experimental tests show that the identification model is effective and efficient: the model can accurately figure out the pollutant amounts or positions no matter single pollution source or multiple sources. Especially when the population size of BGA is set as 10, the computing results are sound agree with analytic results for a single source amount and position identification, the relative errors are no more than 5 %. For cases of multi-point sources and multi-variable, there are some errors in computing results for the reasons that there exist many possible combinations of the pollution sources. But, with the help of previous experience to narrow the search scope, the relative errors of the identification results are less than 5 %, which proves the established source identification model can be used to direct emergency responses.

  10. Parameter Estimation in Rainfall-Runoff Modelling Using Distributed Versions of Particle Swarm Optimization Algorithm

    Directory of Open Access Journals (Sweden)

    Michala Jakubcová

    2015-01-01

    Full Text Available The presented paper provides the analysis of selected versions of the particle swarm optimization (PSO algorithm. The tested versions of the PSO were combined with the shuffling mechanism, which splits the model population into complexes and performs distributed PSO optimization. One of them is a new proposed PSO modification, APartW, which enhances the global exploration and local exploitation in the parametric space during the optimization process through the new updating mechanism applied on the PSO inertia weight. The performances of four selected PSO methods were tested on 11 benchmark optimization problems, which were prepared for the special session on single-objective real-parameter optimization CEC 2005. The results confirm that the tested new APartW PSO variant is comparable with other existing distributed PSO versions, AdaptW and LinTimeVarW. The distributed PSO versions were developed for finding the solution of inverse problems related to the estimation of parameters of hydrological model Bilan. The results of the case study, made on the selected set of 30 catchments obtained from MOPEX database, show that tested distributed PSO versions provide suitable estimates of Bilan model parameters and thus can be used for solving related inverse problems during the calibration process of studied water balance hydrological model.

  11. Designing A Nonlinear Integer Programming Model For A Cross-Dock By A Genetic Algorithm

    Directory of Open Access Journals (Sweden)

    Bahareh Vaisi

    2015-03-01

    Full Text Available Abstract This paper presents a non-linear integer programming model for a cross-dock problem that considers the total transportation cost of inbound and outbound trucks from an origin to a destination and the total cost of assigning strip and stack doors to trucks based on their number of trips and the distance between doors in cross-dock. In previous studies these two cost-based problems are modeled separately however it is more realistic and practical to use both of them as an integrated cross-docking model. Additionally this model is solved for a randomly generated numerical example with three suppliers and two customers by the use of a genetic algorithm. By comparing two different parameter levels i.e. low and high numbers of populations the optimum solution is obtained considering a high level population size. A number of strip and stack doors are equal to a number of inbound and outbound trucks in the same sequence as 4 and 6 respectively. Finally the conclusion is presented.

  12. A face recognition algorithm based on multiple individual discriminative models

    DEFF Research Database (Denmark)

    Fagertun, Jens; Gomez, David Delgado; Ersbøll, Bjarne Kjær

    2005-01-01

    Abstract—In this paper, a novel algorithm for facial recognition is proposed. The technique combines the color texture and geometrical configuration provided by face images. Landmarks and pixel intensities are used by Principal Component Analysis and Fisher Linear Discriminant Analysis to associa...... as an accurate and robust tool for facial identification and unknown detection....

  13. Model-based remote sensing algorithms for particulate organic ...

    Indian Academy of Sciences (India)

    PCA algorithms based on the first three, four, and five modes accounted for 90, 95, and 98% of total variance and yielded significant correlations with POC with 2 = 0.89, 0.92, and 0.93. These full waveband approaches provided robust estimates of POC in various water types. Three different analyses (root mean square ...

  14. An Iterative Algorithm to Determine the Dynamic User Equilibrium in a Traffic Simulation Model

    Science.gov (United States)

    Gawron, C.

    An iterative algorithm to determine the dynamic user equilibrium with respect to link costs defined by a traffic simulation model is presented. Each driver's route choice is modeled by a discrete probability distribution which is used to select a route in the simulation. After each simulation run, the probability distribution is adapted to minimize the travel costs. Although the algorithm does not depend on the simulation model, a queuing model is used for performance reasons. The stability of the algorithm is analyzed for a simple example network. As an application example, a dynamic version of Braess's paradox is studied.

  15. Analytic solution of the Starobinsky model for inflation

    Energy Technology Data Exchange (ETDEWEB)

    Paliathanasis, Andronikos [Universidad Austral de Chile, Instituto de Ciencias Fisicas y Matematicas, Valdivia (Chile); Durban University of Technology, Institute of Systems Science, Durban (South Africa)

    2017-07-15

    We prove that the field equations of the Starobinsky model for inflation in a Friedmann-Lemaitre-Robertson-Walker metric constitute an integrable system. The analytical solution in terms of a Painleve series for the Starobinsky model is presented for the case of zero and nonzero spatial curvature. In both cases the leading-order term describes the radiation era provided by the corresponding higher-order theory. (orig.)

  16. Stochastic models of solute transport in highly heterogeneous geologic media

    Energy Technology Data Exchange (ETDEWEB)

    Semenov, V.N.; Korotkin, I.A.; Pruess, K.; Goloviznin, V.M.; Sorokovikova, O.S.

    2009-09-15

    A stochastic model of anomalous diffusion was developed in which transport occurs by random motion of Brownian particles, described by distribution functions of random displacements with heavy (power-law) tails. One variant of an effective algorithm for random function generation with a power-law asymptotic and arbitrary factor of asymmetry is proposed that is based on the Gnedenko-Levy limit theorem and makes it possible to reproduce all known Levy {alpha}-stable fractal processes. A two-dimensional stochastic random walk algorithm has been developed that approximates anomalous diffusion with streamline-dependent and space-dependent parameters. The motivation for introducing such a type of dispersion model is the observed fact that tracers in natural aquifers spread at different super-Fickian rates in different directions. For this and other important cases, stochastic random walk models are the only known way to solve the so-called multiscaling fractional order diffusion equation with space-dependent parameters. Some comparisons of model results and field experiments are presented.

  17. GASAKe: forecasting landslide activations by a genetic-algorithms based hydrological model

    Science.gov (United States)

    Terranova, O. G.; Gariano, S. L.; Iaquinta, P.; Iovine, G. G. R.

    2015-02-01

    GASAKe is a new hydrological model aimed at forecasting the triggering of landslides. The model is based on genetic-algorithms and allows to obtaining thresholds of landslide activation from the set of historical occurrences and from the rainfall series. GASAKe can be applied to either single landslides or set of similar slope movements in a homogeneous environment. Calibration of the model is based on genetic-algorithms, and provides for families of optimal, discretized solutions (kernels) that maximize the fitness function. Starting from these latter, the corresponding mobility functions (i.e. the predictive tools) can be obtained through convolution with the rain series. The base time of the kernel is related to the magnitude of the considered slope movement, as well as to hydro-geological complexity of the site. Generally, smaller values are expected for shallow slope instabilities with respect to large-scale phenomena. Once validated, the model can be applied to estimate the timing of future landslide activations in the same study area, by employing recorded or forecasted rainfall series. Example of application of GASAKe to a medium-scale slope movement (the Uncino landslide at San Fili, in Calabria, Southern Italy) and to a set of shallow landslides (in the Sorrento Peninsula, Campania, Southern Italy) are discussed. In both cases, a successful calibration of the model has been achieved, despite unavoidable uncertainties concerning the dates of landslide occurrence. In particular, for the Sorrento Peninsula case, a fitness of 0.81 has been obtained by calibrating the model against 10 dates of landslide activation; in the Uncino case, a fitness of 1 (i.e. neither missing nor false alarms) has been achieved against 5 activations. As for temporal validation, the experiments performed by considering the extra dates of landslide activation have also proved satisfactory. In view of early-warning applications for civil protection purposes, the capability of the

  18. Improved belief propagation algorithm finds many Bethe states in the random-field Ising model on random graphs

    Science.gov (United States)

    Perugini, G.; Ricci-Tersenghi, F.

    2018-01-01

    We first present an empirical study of the Belief Propagation (BP) algorithm, when run on the random field Ising model defined on random regular graphs in the zero temperature limit. We introduce the notion of extremal solutions for the BP equations, and we use them to fix a fraction of spins in their ground state configuration. At the phase transition point the fraction of unconstrained spins percolates and their number diverges with the system size. This in turn makes the associated optimization problem highly non trivial in the critical region. Using the bounds on the BP messages provided by the extremal solutions we design a new and very easy to implement BP scheme which is able to output a large number of stable fixed points. On one hand this new algorithm is able to provide the minimum energy configuration with high probability in a competitive time. On the other hand we found that the number of fixed points of the BP algorithm grows with the system size in the critical region. This unexpected feature poses new relevant questions about the physics of this class of models.

  19. Modeling skin collimation using the electron pencil beam redefinition algorithm.

    Science.gov (United States)

    Chi, Pai-Chun M; Hogstrom, Kenneth R; Starkschall, George; Antolak, John A; Boyd, Robert A

    2005-11-01

    Skin collimation is an important tool for electron beam therapy that is used to minimize the penumbra when treating near critical structures, at extended treatment distances, with bolus, or using arc therapy. It is usually made of lead or lead alloy material that conforms to and is placed on patient surface. Presently, commercially available treatment-planning systems lack the ability to model skin collimation and to accurately calculate dose in its presence. The purpose of the present work was to evaluate the use of the pencil beam redefinition algorithm (PBRA) in calculating dose in the presence of skin collimation. Skin collimation was incorporated into the PBRA by terminating the transport of electrons once they enter the skin collimator. Both fixed- and arced-beam dose calculations for arced-beam geometries were evaluated by comparing them with measured dose distributions for 10- and 15-MeV beams. Fixed-beam dose distributions were measured in water at 88-cm source-to-surface distance with an air gap of 32 cm. The 6 x 20-cm2 field (dimensions projected to isocenter) had a 10-mm thick lead collimator placed on the surface of the water with its edge 5 cm inside the field's edge located at +10 cm. Arced-beam dose distributions were measured in a 13.5-cm radius polystyrene circular phantom. The beam was arced 90 degrees (-45 degrees to +45 degrees), and 10-mm thick lead collimation was placed at +/- 30 degrees. For the fixed beam at 10 MeV, the PBRA- calculated dose agreed with measured dose to within 2.0-mm distance to agreement (DTA) in the regions of high-dose gradient and 2.0% in regions of low dose gradient. At 15 MeV, the PBRA agreed to within a 2.0-mm DTA in the regions of high-dose gradient; however, the PBRA underestimated the dose by as much as 5.3% over small regions at depths less than 2 cm because it did not model electrons scattered from the edge of the skin collimation. For arced beams at 10 MeV, the agreement was 1-mm DTA in the high-dose gradient

  20. Study on solitary word based on HMM model and Baum-Welch algorithm

    Directory of Open Access Journals (Sweden)

    Junxia CHEN

    Full Text Available This paper introduces the principle of Hidden Markov Model, which is used to describe the Markov process with unknown parameters, is a probability model to describe the statistical properties of the random process. On this basis, designed a solitary word detection experiment based on HMM model, by optimizing the experimental model, Using Baum-Welch algorithm for training the problem of solving the HMM model, HMM model to estimate the parameters of the λ value is found, in this view of mathematics equivalent to other linear prediction coefficient. This experiment in reducing unnecessary HMM training at the same time, reduced the algorithm complexity. In order to test the effectiveness of the Baum-Welch algorithm, The simulation of experimental data, the results show that the algorithm is effective.

  1. Travelling Wave Solutions in Multigroup Age-Structured Epidemic Models

    Science.gov (United States)

    Ducrot, Arnaut; Magal, Pierre; Ruan, Shigui

    2010-01-01

    Age-structured epidemic models have been used to describe either the age of individuals or the age of infection of certain diseases and to determine how these characteristics affect the outcomes and consequences of epidemiological processes. Most results on age-structured epidemic models focus on the existence, uniqueness, and convergence to disease equilibria of solutions. In this paper we investigate the existence of travelling wave solutions in a deterministic age-structured model describing the circulation of a disease within a population of multigroups. Individuals of each group are able to move with a random walk which is modelled by the classical Fickian diffusion and are classified into two subclasses, susceptible and infective. A susceptible individual in a given group can be crisscross infected by direct contact with infective individuals of possibly any group. This process of transmission can depend upon the age of the disease of infected individuals. The goal of this paper is to provide sufficient conditions that ensure the existence of travelling wave solutions for the age-structured epidemic model. The case of two population groups is numerically investigated which applies to the crisscross transmission of feline immunodeficiency virus (FIV) and some sexual transmission diseases.

  2. SPICE Modeling and Simulation of a MPPT Algorithm

    Directory of Open Access Journals (Sweden)

    Miona Andrejević Stošović

    2014-06-01

    Full Text Available One among several equally important subsystems of a standalone photovoltaic (PV system is the circuit for maximum power point tracking (MPPT. There are several algorithms that may be used for it. In this paper we choose such an algorithm based on the maximum simplicity criteria. Then we make some small modifications to it in order to make it more robust. We synthesize a circuit built out of elements from the list of elements recognized by SPICE. The inputs are the voltage and the current at the PV panel to DC-DC converter interface. Its task is to generate a pulse width modulated pulse train whose duty ratio is defined to keep the input impedance of the DC-DC converter at the optimal value.

  3. Event-chain algorithm for the Heisenberg model: Evidence for z≃1 dynamic scaling.

    Science.gov (United States)

    Nishikawa, Yoshihiko; Michel, Manon; Krauth, Werner; Hukushima, Koji

    2015-12-01

    We apply the event-chain Monte Carlo algorithm to the three-dimensional ferromagnetic Heisenberg model. The algorithm is rejection-free and also realizes an irreversible Markov chain that satisfies global balance. The autocorrelation functions of the magnetic susceptibility and the energy indicate a dynamical critical exponent z≈1 at the critical temperature, while that of the magnetization does not measure the performance of the algorithm. We show that the event-chain Monte Carlo algorithm substantially reduces the dynamical critical exponent from the conventional value of z≃2.

  4. Atmosphere Clouds Model Algorithm for Solving Optimal Reactive Power Dispatch Problem

    Directory of Open Access Journals (Sweden)

    Lenin Kanagasabai

    2014-04-01

    Full Text Available In this paper, a new method, called Atmosphere Clouds Model (ACM algorithm, used for solving optimal reactive power dispatch problem. ACM stochastic optimization algorithm stimulated from the behavior of cloud in the natural earth. ACM replicate the generation behavior, shift behavior and extend behavior of cloud. The projected (ACM algorithm has been tested on standard IEEE 30 bus test system and simulation results shows clearly about the superior performance of the proposed algorithm in plummeting the real power loss. Normal 0 false false false EN-IN X-NONE X-NONE

  5. Covariance Structure Model Fit Testing under Missing Data: An Application of the Supplemented EM Algorithm

    Science.gov (United States)

    Cai, Li; Lee, Taehun

    2009-01-01

    We apply the Supplemented EM algorithm (Meng & Rubin, 1991) to address a chronic problem with the "two-stage" fitting of covariance structure models in the presence of ignorable missing data: the lack of an asymptotically chi-square distributed goodness-of-fit statistic. We show that the Supplemented EM algorithm provides a…

  6. Inferring the structure of latent class models using a genetic algorithm

    NARCIS (Netherlands)

    van der Maas, H.L.J.; Raijmakers, M.E.J.; Visser, I.

    2005-01-01

    Present optimization techniques in latent class analysis apply the expectation maximization algorithm or the Newton-Raphson algorithm for optimizing the parameter values of a prespecified model. These techniques can be used to find maximum likelihood estimates of the parameters, given the specified

  7. Global Convergence of the EM Algorithm for Unconstrained Latent Variable Models with Categorical Indicators

    Science.gov (United States)

    Weissman, Alexander

    2013-01-01

    Convergence of the expectation-maximization (EM) algorithm to a global optimum of the marginal log likelihood function for unconstrained latent variable models with categorical indicators is presented. The sufficient conditions under which global convergence of the EM algorithm is attainable are provided in an information-theoretic context by…

  8. Automated Test Assembly for Cognitive Diagnosis Models Using a Genetic Algorithm

    Science.gov (United States)

    Finkelman, Matthew; Kim, Wonsuk; Roussos, Louis A.

    2009-01-01

    Much recent psychometric literature has focused on cognitive diagnosis models (CDMs), a promising class of instruments used to measure the strengths and weaknesses of examinees. This article introduces a genetic algorithm to perform automated test assembly alongside CDMs. The algorithm is flexible in that it can be applied whether the goal is to…

  9. An approach based on genetic algorithms with coding in real for the solution of a DC OPF to hydrothermal systems; Uma abordagem baseada em algoritmos geneticos com codificacao em real para a solucao de um FPO DC para sistemas hidrotermicos

    Energy Technology Data Exchange (ETDEWEB)

    Barbosa, Diego R.; Silva, Alessandro L. da; Luciano, Edson Jose Rezende; Nepomuceno, Leonardo [Universidade Estadual Paulista (UNESP), Bauru, SP (Brazil). Dept. de Engenharia Eletrica], Emails: diego_eng.eletricista@hotmail.com, alessandrolopessilva@uol.com.br, edson.joserl@uol.com.br, leo@feb.unesp.br

    2009-07-01

    Problems of DC Optimal Power Flow (OPF) have been solved by various conventional optimization methods. When the modeling of DC OPF involves discontinuous functions or not differentiable, the use of solution methods based on conventional optimization is often not possible because of the difficulty in calculating the gradient vectors at points of discontinuity/non-differentiability of these functions. This paper proposes a method for solving the DC OPF based on Genetic Algorithms (GA) with real coding. The proposed GA has specific genetic operators to improve the quality and viability of the solution. The results are analyzed for an IEEE test system, and its solutions are compared, when possible, with those obtained by a method of interior point primal-dual logarithmic barrier. The results highlight the robustness of the method and feasibility of obtaining the solution to real systems.

  10. Efficient cache oblivious algorithms for randomized divide-and-conquer on the multicore model

    OpenAIRE

    Sharma, Neeraj; Sen, Sandeep

    2012-01-01

    In this paper we present randomized algorithms for sorting and convex hull that achieves optimal performance (for speed-up and cache misses) on the multicore model with private cache model. Our algorithms are cache oblivious and generalize the randomized divide and conquer strategy given by Reischuk and Reif and Sen. Although the approach yielded optimal speed-up in the PRAM model, we require additional techniques to optimize cache-misses in an oblivious setting. Under a mild assumption on in...

  11. Mesoscale modeling of solute precipitation and radiation damage

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Yongfeng [Idaho National Lab. (INL), Idaho Falls, ID (United States); Schwen, Daniel [Idaho National Lab. (INL), Idaho Falls, ID (United States); Ke, Huibin [Idaho National Lab. (INL), Idaho Falls, ID (United States); Univ. of Wisconsin, Madison, WI (United States); Bai, Xianming [Idaho National Lab. (INL), Idaho Falls, ID (United States); Hales, Jason [Idaho National Lab. (INL), Idaho Falls, ID (United States)

    2015-09-01

    This report summarizes the low length scale effort during FY 2014 in developing mesoscale capabilities for microstructure evolution in reactor pressure vessels. During operation, reactor pressure vessels are subject to hardening and embrittlement caused by irradiation-induced defect accumulation and irradiation-enhanced solute precipitation. Both defect production and solute precipitation start from the atomic scale, and manifest their eventual effects as degradation in engineering-scale properties. To predict the property degradation, multiscale modeling and simulation are needed to deal with the microstructure evolution, and to link the microstructure feature to material properties. In this report, the development of mesoscale capabilities for defect accumulation and solute precipitation are summarized. Atomic-scale efforts that supply information for the mesoscale capabilities are also included.

  12. Ruthenium speciation in model nuclear fuel process solutions

    International Nuclear Information System (INIS)

    Koster, Anne L.; May, Iain; Sharrad, Clint A.; Wright, Des; Owens, Ivan F.; Charnock, John M.; Hennig, Christoph

    2004-01-01

    Ru speciation is being investigated systematically from models of high level waste solutions right through to the calcination process and the vitrified glass product. The characterisation of these species is complicated due to the fact that a wide range of ruthenium nitrosyl/nitrite/nitrate complexes can be present in nitric acid waste solutions. The general formula for these complexes is RuNO(NO 3 ) x (NO 2 ) y (OH) z (H 2 O) 5-x-y-z +3-x-y-z . A range of different techniques has been used for the characterisation of these species in solution, including electron absorption spectroscopy, vibrational spectroscopy, multi-nuclear magnetic resonance spectroscopy and X-ray absorption spectroscopy. (authors)

  13. Multiple-try differential evolution adaptive Metropolis for efficient solution of highly parameterized models

    Science.gov (United States)

    Eric, L.; Vrugt, J. A.

    2010-12-01

    Spatially distributed hydrologic models potentially contain hundreds of parameters that need to be derived by calibration against a historical record of input-output data. The quality of this calibration strongly determines the predictive capability of the model and thus its usefulness for science-based decision making and forecasting. Unfortunately, high-dimensional optimization problems are typically difficult to solve. Here we present our recent developments to the Differential Evolution Adaptive Metropolis (DREAM) algorithm (Vrugt et al., 2009) to warrant efficient solution of high-dimensional parameter estimation problems. The algorithm samples from an archive of past states (Ter Braak and Vrugt, 2008), and uses multiple-try Metropolis sampling (Liu et al., 2000) to decrease the required burn-in time for each individual chain and increase efficiency of posterior sampling. This approach is hereafter referred to as MT-DREAM. We present results for 2 synthetic mathematical case studies, and 2 real-world examples involving from 10 to 240 parameters. Results for those cases show that our multiple-try sampler, MT-DREAM, can consistently find better solutions than other Bayesian MCMC methods. Moreover, MT-DREAM is admirably suited to be implemented and ran on a parallel machine and is therefore a powerful method for posterior inference.

  14. Model-Based Fault Diagnosis Techniques Design Schemes, Algorithms and Tools

    CERN Document Server

    Ding, Steven X

    2013-01-01

    Guaranteeing a high system performance over a wide operating range is an important issue surrounding the design of automatic control systems with successively increasing complexity. As a key technology in the search for a solution, advanced fault detection and identification (FDI) is receiving considerable attention. This book introduces basic model-based FDI schemes, advanced analysis and design algorithms, and mathematical and control-theoretic tools. This second edition of Model-Based Fault Diagnosis Techniques contains: ·         new material on fault isolation and identification, and fault detection in feedback control loops; ·         extended and revised treatment of systematic threshold determination for systems with both deterministic unknown inputs and stochastic noises; addition of the continuously-stirred tank heater as a representative process-industrial benchmark; and ·         enhanced discussion of residual evaluation in stochastic processes. Model-based Fault Diagno...

  15. Solution Approach to Automatic Generation Control Problem Using Hybridized Gravitational Search Algorithm Optimized PID and FOPID Controllers

    Directory of Open Access Journals (Sweden)

    DAHIYA, P.

    2015-05-01

    Full Text Available This paper presents the application of hybrid opposition based disruption operator in gravitational search algorithm (DOGSA to solve automatic generation control (AGC problem of four area hydro-thermal-gas interconnected power system. The proposed DOGSA approach combines the advantages of opposition based learning which enhances the speed of convergence and disruption operator which has the ability to further explore and exploit the search space of standard gravitational search algorithm (GSA. The addition of these two concepts to GSA increases its flexibility for solving the complex optimization problems. This paper addresses the design and performance analysis of DOGSA based proportional integral derivative (PID and fractional order proportional integral derivative (FOPID controllers for automatic generation control problem. The proposed approaches are demonstrated by comparing the results with the standard GSA, opposition learning based GSA (OGSA and disruption based GSA (DGSA. The sensitivity analysis is also carried out to study the robustness of DOGSA tuned controllers in order to accommodate variations in operating load conditions, tie-line synchronizing coefficient, time constants of governor and turbine. Further, the approaches are extended to a more realistic power system model by considering the physical constraints such as thermal turbine generation rate constraint, speed governor dead band and time delay.

  16. On-line monitoring the extract process of Fu-fang Shuanghua oral solution using near infrared spectroscopy and different PLS algorithms

    Science.gov (United States)

    Kang, Qian; Ru, Qingguo; Liu, Yan; Xu, Lingyan; Liu, Jia; Wang, Yifei; Zhang, Yewen; Li, Hui; Zhang, Qing; Wu, Qing

    2016-01-01

    An on-line near infrared (NIR) spectroscopy monitoring method with an appropriate multivariate calibration method was developed for the extraction process of Fu-fang Shuanghua oral solution (FSOS). On-line NIR spectra were collected through two fiber optic probes, which were designed to transmit NIR radiation by a 2 mm flange. Partial least squares (PLS), interval PLS (iPLS) and synergy interval PLS (siPLS) algorithms were used comparatively for building the calibration regression models. During the extraction process, the feasibility of NIR spectroscopy was employed to determine the concentrations of chlorogenic acid (CA) content, total phenolic acids contents (TPC), total flavonoids contents (TFC) and soluble solid contents (SSC). High performance liquid chromatography (HPLC), ultraviolet spectrophotometric method (UV) and loss on drying methods were employed as reference methods. Experiment results showed that the performance of siPLS model is the best compared with PLS and iPLS. The calibration models for AC, TPC, TFC and SSC had high values of determination coefficients of (R2) (0.9948, 0.9992, 0.9950 and 0.9832) and low root mean square error of cross validation (RMSECV) (0.0113, 0.0341, 0.1787 and 1.2158), which indicate a good correlation between reference values and NIR predicted values. The overall results show that the on line detection method could be feasible in real application and would be of great value for monitoring the mixed decoction process of FSOS and other Chinese patent medicines.

  17. Selection of robust variables for transfer of classification models employing the successive projections algorithm.

    Science.gov (United States)

    Milanez, Karla Danielle Tavares Melo; Araújo Nóbrega, Thiago César; Silva Nascimento, Danielle; Galvão, Roberto Kawakami Harrop; Pontes, Márcio José Coelho

    2017-09-01

    Multivariate models have been widely used in analytical problems involving quantitative and qualitative analyzes. However, there are cases in which a model is not applicable to spectra of samples obtained under new experimental conditions or in an instrument not involved in the modeling step. A solution to this problem is the transfer of multivariate models, usually performed using standardization of the spectral responses or enhancement of the robustness of the model. This present paper proposes two new criteria for selection of robust variables for classification transfer employing the successive projections algorithm (SPA). These variables are then used to build models based on linear discriminant analysis (LDA) with low sensitivity with respect to the differences between the responses of the instruments involved. For this purpose, transfer samples are included in the calculation of the cost for each subset of variables under consideration. The proposed methods are evaluated for two case studies involving identification of adulteration of extra virgin olive oil (EVOO) and hydrated ethyl alcohol fuel (HEAF) using UV-Vis and NIR spectroscopy, respectively. In both cases, similar or better classification transfer results (obtained for a test set measured on the secondary instrument) employing the two criteria were obtained in comparison with direct standardization (DS) and piecewise direct standardization (PDS). For the UV-Vis data, both proposed criteria achieved the correct classification rate (CCR) of 85%, while the best CCR obtained for the standardization methods was 81% for DS. For the NIR data, 92.5% of CCR was obtained by both criteria as well as DS. The results demonstrated the possibility of using either of the criteria proposed for building robust models as an alternative to the standardization of spectral responses for transfer of classification. Copyright © 2017 Elsevier B.V. All rights reserved.

  18. Frequency-domain optical tomographic image reconstruction algorithm with the simplified spherical harmonics (SP3) light propagation model.

    Science.gov (United States)

    Kim, Hyun Keol; Montejo, Ludguier D; Jia, Jingfei; Hielscher, Andreas H

    2017-06-01

    We introduce here the finite volume formulation of the frequency-domain simplified spherical harmonics model with n -th order absorption coefficients (FD-SP N ) that approximates the frequency-domain equation of radiative transfer (FD-ERT). We then present the FD-SP N based reconstruction algorithm that recovers absorption and scattering coefficients in biological tissue. The FD-SP N model with 3 rd order absorption coefficient (i.e., FD-SP 3 ) is used as a forward model to solve the inverse problem. The FD-SP 3 is discretized with a node-centered finite volume scheme and solved with a restarted generalized minimum residual (GMRES) algorithm. The absorption and scattering coefficients are retrieved using a limited-memory Broyden-Fletcher-Goldfarb-Shanno (L-BFGS) algorithm. Finally, the forward and inverse algorithms are evaluated using numerical phantoms with optical properties and size that mimic small-volume tissue such as finger joints and small animals. The forward results show that the FD-SP 3 model approximates the FD-ERT (S 12 ) solution within relatively high accuracy; the average error in the phase (<3.7%) and the amplitude (<7.1%) of the partial current at the boundary are reported. From the inverse results we find that the absorption and scattering coefficient maps are more accurately reconstructed with the SP 3 model than those with the SP 1 model. Therefore, this work shows that the FD-SP 3 is an efficient model for optical tomographic imaging of small-volume media with non-diffuse properties both in terms of computational time and accuracy as it requires significantly lower CPU time than the FD-ERT (S 12 ) and also it is more accurate than the FD-SP 1 .

  19. MOESHA: A genetic algorithm for automatic calibration and estimation of parameter uncertainty and sensitivity of hydrologic models

    Science.gov (United States)

    Characterization of uncertainty and sensitivity of model parameters is an essential and often overlooked facet of hydrological modeling. This paper introduces an algorithm called MOESHA that combines input parameter sensitivity analyses with a genetic algorithm calibration routin...

  20. Optimal parallel algorithms for problems modeled by a family of intervals

    Science.gov (United States)

    Olariu, Stephan; Schwing, James L.; Zhang, Jingyuan

    1992-01-01

    A family of intervals on the real line provides a natural model for a vast number of scheduling and VLSI problems. Recently, a number of parallel algorithms to solve a variety of practical problems on such a family of intervals have been proposed in the literature. Computational tools are developed, and it is shown how they can be used for the purpose of devising cost-optimal parallel algorithms for a number of interval-related problems including finding a largest subset of pairwise nonoverlapping intervals, a minimum dominating subset of intervals, along with algorithms to compute the shortest path between a pair of intervals and, based on the shortest path, a parallel algorithm to find the center of the family of intervals. More precisely, with an arbitrary family of n intervals as input, all algorithms run in O(log n) time using O(n) processors in the EREW-PRAM model of computation.

  1. TWO-STEP ALGORITHM OF TRAINING INITIALIZATION FOR ACOUSTIC MODELS BASED ON DEEP NEURAL NETWORKS

    Directory of Open Access Journals (Sweden)

    I. P. Medennikov

    2016-03-01

    Full Text Available This paper presents a two-step initialization algorithm for training of acoustic models based on deep neural networks. The algorithm is focused on reducing the impact of the non-speech segments on the acoustic model training. The idea of the proposed algorithm is to reduce the percentage of non-speech examples in the training set. Effectiveness evaluation of the algorithm has been carried out on the example of English spontaneous telephone speech recognition (Switchboard. The application of the proposed algorithm has led to 3% relative word error rate reduction, compared with the training initialization by restricted Boltzmann machines. The results presented in the paper can be applied in the development of automatic speech recognition systems.

  2. Behavioral Modeling for Mental Health using Machine Learning Algorithms.

    Science.gov (United States)

    Srividya, M; Mohanavalli, S; Bhalaji, N

    2018-04-03

    Mental health is an indicator of emotional, psychological and social well-being of an individual. It determines how an individual thinks, feels and handle situations. Positive mental health helps one to work productively and realize their full potential. Mental health is important at every stage of life, from childhood and adolescence through adulthood. Many factors contribute to mental health problems which lead to mental illness like stress, social anxiety, depression, obsessive compulsive disorder, drug addiction, and personality disorders. It is becoming increasingly important to determine the onset of the mental illness to maintain proper life balance. The nature of machine learning algorithms and Artificial Intelligence (AI) can be fully harnessed for predicting the onset of mental illness. Such applications when implemented in real time will benefit the society by serving as a monitoring tool for individuals with deviant behavior. This research work proposes to apply various machine learning algorithms such as support vector machines, decision trees, naïve bayes classifier, K-nearest neighbor classifier and logistic regression to identify state of mental health in a target group. The responses obtained from the target group for the designed questionnaire were first subject to unsupervised learning techniques. The labels obtained as a result of clustering were validated by computing the Mean Opinion Score. These cluster labels were then used to build classifiers to predict the mental health of an individual. Population from various groups like high school students, college students and working professionals were considered as target groups. The research presents an analysis of applying the aforementioned machine learning algorithms on the target groups and also suggests directions for future work.

  3. Use of the AIC with the EM algorithm: A demonstration of a probability model selection technique

    Energy Technology Data Exchange (ETDEWEB)

    Glosup, J.G.; Axelrod M.C. [Lawrence Livermore National Lab., CA (United States)

    1994-11-15

    The problem of discriminating between two potential probability models, a Gaussian distribution and a mixture of Gaussian distributions, is considered. The focus of our interest is a case where the models are potentially non-nested and the parameters of the mixture model are estimated through the EM algorithm. The AIC, which is frequently used as a criterion for discriminating between non-nested models, is modified to work with the EM algorithm and is shown to provide a model selection tool for this situation. A particular problem involving an infinite mixture distribution known as Middleton`s Class A model is used to demonstrate the effectiveness and limitations of this method.

  4. Solution-focused therapy. Counseling model for busy family physicians.

    OpenAIRE

    Greenberg, G.; Ganshorn, K.; Danilkewich, A.

    2001-01-01

    OBJECTIVE: To provide family doctors in busy office practices with a model for counseling compatible with patient-centred medicine, including the techniques, strategies, and questions necessary for implementation. QUALITY OF EVIDENCE: The MEDLINE database was searched from 1984 to 1999 using the terms psychotherapy in family practice, brief therapy in family practice, solution-focused therapy, and brief psychotherapy. A total of 170 relevant articles were identified; 75 abstracts were retriev...

  5. An improved algorithm to convert CAD model to MCNP geometry model based on STEP file

    International Nuclear Information System (INIS)

    Zhou, Qingguo; Yang, Jiaming; Wu, Jiong; Tian, Yanshan; Wang, Junqiong; Jiang, Hai; Li, Kuan-Ching

    2015-01-01

    Highlights: • Fully exploits common features of cells, making the processing efficient. • Accurately provide the cell position. • Flexible to add new parameters in the structure. • Application of novel structure in INP file processing, conveniently evaluate cell location. - Abstract: MCNP (Monte Carlo N-Particle Transport Code) is a general-purpose Monte Carlo N-Particle code that can be used for neutron, photon, electron, or coupled neutron/photon/electron transport. Its input file, the INP file, has the characteristics of complicated form and is error-prone when describing geometric models. Due to this, a conversion algorithm that can solve the problem by converting general geometric model to MCNP model during MCNP aided modeling is highly needed. In this paper, we revised and incorporated a number of improvements over our previous work (Yang et al., 2013), which was proposed and targeted after STEP file and INP file were analyzed. Results of experiments show that the revised algorithm is more applicable and efficient than previous work, with the optimized extraction of geometry and topology information of the STEP file, as well as the production efficiency of output INP file. This proposed research is promising, and serves as valuable reference for the majority of researchers involved with MCNP-related researches

  6. Reduced-order modeling (ROM) for simulation and optimization powerful algorithms as key enablers for scientific computing

    CERN Document Server

    Milde, Anja; Volkwein, Stefan

    2018-01-01

    This edited monograph collects research contributions and addresses the advancement of efficient numerical procedures in the area of model order reduction (MOR) for simulation, optimization and control. The topical scope includes, but is not limited to, new out-of-the-box algorithmic solutions for scientific computing, e.g. reduced basis methods for industrial problems and MOR approaches for electrochemical processes. The target audience comprises research experts and practitioners in the field of simulation, optimization and control, but the book may also be beneficial for graduate students alike. .

  7. An Empirical Study of Wrappers for Feature Subset Selection based on a Parallel Genetic Algorithm: The Multi-Wrapper Model

    KAUST Repository

    Soufan, Othman

    2012-09-01

    Feature selection is the first task of any learning approach that is applied in major fields of biomedical, bioinformatics, robotics, natural language processing and social networking. In feature subset selection problem, a search methodology with a proper criterion seeks to find the best subset of features describing data (relevance) and achieving better performance (optimality). Wrapper approaches are feature selection methods which are wrapped around a classification algorithm and use a performance measure to select the best subset of features. We analyze the proper design of the objective function for the wrapper approach and highlight an objective based on several classification algorithms. We compare the wrapper approaches to different feature selection methods based on distance and information based criteria. Significant improvement in performance, computational time, and selection of minimally sized feature subsets is achieved by combining different objectives for the wrapper model. In addition, considering various classification methods in the feature selection process could lead to a global solution of desirable characteristics.

  8. Improved Expectation Maximization Algorithm for Gaussian Mixed Model Using the Kernel Method

    Directory of Open Access Journals (Sweden)

    Mohd Izhan Mohd Yusoff

    2013-01-01

    Full Text Available Fraud activities have contributed to heavy losses suffered by telecommunication companies. In this paper, we attempt to use Gaussian mixed model, which is a probabilistic model normally used in speech recognition to identify fraud calls in the telecommunication industry. We look at several issues encountered when calculating the maximum likelihood estimates of the Gaussian mixed model using an Expectation Maximization algorithm. Firstly, we look at a mechanism for the determination of the initial number of Gaussian components and the choice of the initial values of the algorithm using the kernel method. We show via simulation that the technique improves the performance of the algorithm. Secondly, we developed a procedure for determining the order of the Gaussian mixed model using the log-likelihood function and the Akaike information criteria. Finally, for illustration, we apply the improved algorithm to real telecommunication data. The modified method will pave the way to introduce a comprehensive method for detecting fraud calls in future work.

  9. A Multiple Time-Step Finite State Projection Algorithm for the Solution to the Chemical Master Equation

    Science.gov (United States)

    2006-11-30

    begins in state k, the initial probability distribution for the CME was written, pi(0) = δik, where δik is the Kronecker delta . Suppose now that the...initial distribution is given not by the Kronecker delta but by a vector with many non-zero elements. For example, suppose that the initial distribution is...pap-pili epigenetic switch,” Proc. FOSBE , pp. 145–148, August 2005. [16] B. Munsky and M. Khammash, “A reduced model solution for the chemical master

  10. Estimating model error covariances in nonlinear state-space models using Kalman smoothing and the expectation-maximisation algorithm

    KAUST Repository

    Dreano, Denis

    2017-04-05

    Specification and tuning of errors from dynamical models are important issues in data assimilation. In this work, we propose an iterative expectation-maximisation (EM) algorithm to estimate the model error covariances using classical extended and ensemble versions of the Kalman smoother. We show that, for additive model errors, the estimate of the error covariance converges. We also investigate other forms of model error, such as parametric or multiplicative errors. We show that additive Gaussian model error is able to compensate for non additive sources of error in the algorithms we propose. We also demonstrate the limitations of the extended version of the algorithm and recommend the use of the more robust and flexible ensemble version. This article is a proof of concept of the methodology with the Lorenz-63 attractor. We developed an open-source Python library to enable future users to apply the algorithm to their own nonlinear dynamical models.

  11. State-space models - from the EM algorithm to a gradient approach

    DEFF Research Database (Denmark)

    Olsson, Rasmus Kongsgaard; Petersen, Kaare Brandt; Lehn-Schiøler, Tue

    2007-01-01

    Slow convergence is observed in the EM algorithm for linear state-space models. We propose to circumvent the problem by applying any off-the-shelf quasi-Newton-type optimizer, which operates on the gradient of the log-likelihood function. Such an algorithm is a practical alternative due to the fact...... that the exact gradient of the log-likelihood function can be computed by recycling components of the expectation-maximization (EM) algorithm. We demonstrate the efficiency of the proposed method in three relevant instances of the linear state-space model. In high signal-to-noise ratios, where EM is particularly...

  12. Filtering Based Recursive Least Squares Algorithm for Multi-Input Multioutput Hammerstein Models

    Directory of Open Access Journals (Sweden)

    Ziyun Wang

    2014-01-01

    Full Text Available This paper considers the parameter estimation problem for Hammerstein multi-input multioutput finite impulse response (FIR-MA systems. Filtered by the noise transfer function, the FIR-MA model is transformed into a controlled autoregressive model. The key-term variable separation principle is used to derive a data filtering based recursive least squares algorithm. The numerical examples confirm that the proposed algorithm can estimate parameters more accurately and has a higher computational efficiency compared with the recursive least squares algorithm.

  13. Filtering Based Recursive Least Squares Algorithm for Multi-Input Multioutput Hammerstein Models

    OpenAIRE

    Wang, Ziyun; Wang, Yan; Ji, Zhicheng

    2014-01-01

    This paper considers the parameter estimation problem for Hammerstein multi-input multioutput finite impulse response (FIR-MA) systems. Filtered by the noise transfer function, the FIR-MA model is transformed into a controlled autoregressive model. The key-term variable separation principle is used to derive a data filtering based recursive least squares algorithm. The numerical examples confirm that the proposed algorithm can estimate parameters more accurately and has a higher computational...

  14. Asymptotically exact solution of a local copper-oxide model

    International Nuclear Information System (INIS)

    Zhang Guangming; Yu Lu.

    1994-03-01

    We present an asymptotically exact solution of a local copper-oxide model abstracted from the multi-band models. The phase diagram is obtained through the renormalization-group analysis of the partition function. In the strong coupling regime, we find an exactly solved line, which crosses the quantum critical point of the mixed valence regime separating two different Fermi-liquid (FL) phases. At this critical point, a many-particle resonance is formed near the chemical potential, and a marginal-FL spectrum can be derived for the spin and charge susceptibilities. (author). 15 refs, 1 fig

  15. Exact solutions for the spin tune for model storage rings

    CERN Document Server

    Mane, S R

    2002-01-01

    We present exact analytical expressions for the spin tune for arbitrary values of the orbital action for several storage ring models. The models we treat contain Siberian Snakes, the use of which is essential to preserve the polarization of beams in high-energy proton storage rings. Our solutions contain some novel features. We also prove a previously conjectured claim about the behavior of spin tuneshifts in rings with multiple Snakes. The conjecture is based on numerical simulations, but our proof is analytical, and also nonperturbative.

  16. Bessel collocation approach for approximate solutions of Hantavirus infection model

    Directory of Open Access Journals (Sweden)

    Suayip Yuzbasi

    2017-11-01

    Full Text Available In this study, a collocation method is introduced to find the approximate solutions of Hantavirus infection model which is a system of nonlinear ordinary differential equations. The method is based on the Bessel functions of the first kind, matrix operations and collocation points. This method converts Hantavirus infection model into a matrix equation in terms of the Bessel functions of first kind, matrix operations and collocation points. The matrix equation corresponds to a system of nonlinear equations with the unknown Bessel coefficients. The reliability and efficiency of the suggested scheme are demonstrated by numerical applications and all numerical calculations have been done by using a program written in Maple.

  17. Asymptotic solutions of diffusion models for risk reserves

    Directory of Open Access Journals (Sweden)

    S. Shao

    2003-01-01

    Full Text Available We study a family of diffusion models for risk reserves which account for the investment income earned and for the inflation experienced on claim amounts. After we defined the process of the conditional probability of ruin over finite time and imposed the appropriate boundary conditions, classical results from the theory of diffusion processes turn the stochastic differential equation to a special class of initial and boundary value problems defined by a linear diffusion equation. Armed with asymptotic analysis and perturbation theory, we obtain the asymptotic solutions of the diffusion models (possibly degenerate governing the conditional probability of ruin over a finite time in terms of interest rate.

  18. Spectroscopic properties of vitamin E models in solution

    Science.gov (United States)

    Oliveira, L. B. A.; Colherinhas, G.; Fonseca, T. L.; Castro, M. A.

    2015-05-01

    We investigate the first absorption band and the 13C and 17O magnetic shieldings of vitamin E models in chloroform and in water using the S-MC/QM methodology in combination with the TD-DFT and GIAO approaches. The results show that the solvent effects on these spectroscopic properties are small but a proper description of the solvent shift for 17O magnetic shielding of the hydroxyl group in water requires the use of explicit solute-solvent hydrogen bonds. In addition, the effect of the replacement of hydrogen atoms by methyl groups in the vitamin E models only affects magnetic shieldings.

  19. Microfluidic model experiments on the injectability of monoclonal antibody solutions

    Science.gov (United States)

    Duchene, Charles; Filipe, Vasco; Nakach, Mostafa; Huille, Sylvain; Lindner, Anke

    2017-11-01

    Autoinjection devices that allow patients to self-administer medicine are becoming used more frequently; however, this advance comes with an increased need for precision in the injection process. The rare occurrence of protein aggregates in solutions of monoclonal antibodies constitutes a threat to the reliability of such devices. Here we study the flow of protein solutions containing aggregates in microfluidic model systems, mimicking injection devices, to gain fundamental understanding of the catastrophic clogging of constrictions of given size. We form aggregates by mechanically shaking or heating antibody solutions and then inject these solutions into microfluidic channels with varying types of constrictions. Geometrical clogging occurs when aggregates reach the size of the constriction and can in some cases be undone by increasing the applied pressure. We perform systematic experiments varying the relative aggregate size and the flow rate or applied pressure. The mechanical deformation of aggregates during their passage through constrictions is investigated to gain a better understanding of the clogging and unclogging mechanisms.

  20. Prediction models and control algorithms for predictive applications of setback temperature in cooling systems

    International Nuclear Information System (INIS)

    Moon, Jin Woo; Yoon, Younju; Jeon, Young-Hoon; Kim, Sooyoung

    2017-01-01

    Highlights: • Initial ANN model was developed for predicting the time to the setback temperature. • Initial model was optimized for producing accurate output. • Optimized model proved its prediction accuracy. • ANN-based algorithms were developed and tested their performance. • ANN-based algorithms presented superior thermal comfort or energy efficiency. - Abstract: In this study, a temperature control algorithm was developed to apply a setback temperature predictively for the cooling system of a residential building during occupied periods by residents. An artificial neural network (ANN) model was developed to determine the required time for increasing the current indoor temperature to the setback temperature. This study involved three phases: development of the initial ANN-based prediction model, optimization and testing of the initial model, and development and testing of three control algorithms. The development and performance testing of the model and algorithm were conducted using TRNSYS and MATLAB. Through the development and optimization process, the final ANN model employed indoor temperature and the temperature difference between the current and target setback temperature as two input neurons. The optimal number of hidden layers, number of neurons, learning rate, and moment were determined to be 4, 9, 0.6, and 0.9, respectively. The tangent–sigmoid and pure-linear transfer function was used in the hidden and output neurons, respectively. The ANN model used 100 training data sets with sliding-window method for data management. Levenberg-Marquart training method was employed for model training. The optimized model had a prediction accuracy of 0.9097 root mean square errors when compared with the simulated results. Employing the ANN model, ANN-based algorithms maintained indoor temperatures better within target ranges. Compared to the conventional algorithm, the ANN-based algorithms reduced the duration of time, in which the indoor temperature