A Globally Convergent Parallel SSLE Algorithm for Inequality Constrained Optimization
Directory of Open Access Journals (Sweden)
Zhijun Luo
2014-01-01
Full Text Available A new parallel variable distribution algorithm based on interior point SSLE algorithm is proposed for solving inequality constrained optimization problems under the condition that the constraints are block-separable by the technology of sequential system of linear equation. Each iteration of this algorithm only needs to solve three systems of linear equations with the same coefficient matrix to obtain the descent direction. Furthermore, under certain conditions, the global convergence is achieved.
Evolutionary constrained optimization
Deb, Kalyanmoy
2015-01-01
This book makes available a self-contained collection of modern research addressing the general constrained optimization problems using evolutionary algorithms. Broadly the topics covered include constraint handling for single and multi-objective optimizations; penalty function based methodology; multi-objective based methodology; new constraint handling mechanism; hybrid methodology; scaling issues in constrained optimization; design of scalable test problems; parameter adaptation in constrained optimization; handling of integer, discrete and mix variables in addition to continuous variables; application of constraint handling techniques to real-world problems; and constrained optimization in dynamic environment. There is also a separate chapter on hybrid optimization, which is gaining lots of popularity nowadays due to its capability of bridging the gap between evolutionary and classical optimization. The material in the book is useful to researchers, novice, and experts alike. The book will also be useful...
Order-Constrained Solutions in K-Means Clustering: Even Better than Being Globally Optimal
Steinley, Douglas; Hubert, Lawrence
2008-01-01
This paper proposes an order-constrained K-means cluster analysis strategy, and implements that strategy through an auxiliary quadratic assignment optimization heuristic that identifies an initial object order. A subsequent dynamic programming recursion is applied to optimally subdivide the object set subject to the order constraint. We show that…
Linearly constrained minimax optimization
DEFF Research Database (Denmark)
Madsen, Kaj; Schjær-Jacobsen, Hans
1978-01-01
We present an algorithm for nonlinear minimax optimization subject to linear equality and inequality constraints which requires first order partial derivatives. The algorithm is based on successive linear approximations to the functions defining the problem. The resulting linear subproblems...
Ekren, Ibrahim; Soner, H. Mete
2018-03-01
The classical duality theory of Kantorovich (C R (Doklady) Acad Sci URSS (NS) 37:199-201, 1942) and Kellerer (Z Wahrsch Verw Gebiete 67(4):399-432, 1984) for classical optimal transport is generalized to an abstract framework and a characterization of the dual elements is provided. This abstract generalization is set in a Banach lattice X with an order unit. The problem is given as the supremum over a convex subset of the positive unit sphere of the topological dual of X and the dual problem is defined on the bi-dual of X. These results are then applied to several extensions of the classical optimal transport.
Slope constrained Topology Optimization
DEFF Research Database (Denmark)
Petersson, J.; Sigmund, Ole
1998-01-01
The problem of minimum compliance topology optimization of an elastic continuum is considered. A general continuous density-energy relation is assumed, including variable thickness sheet models and artificial power laws. To ensure existence of solutions, the design set is restricted by enforcing...
Trends in PDE constrained optimization
Benner, Peter; Engell, Sebastian; Griewank, Andreas; Harbrecht, Helmut; Hinze, Michael; Rannacher, Rolf; Ulbrich, Stefan
2014-01-01
Optimization problems subject to constraints governed by partial differential equations (PDEs) are among the most challenging problems in the context of industrial, economical and medical applications. Almost the entire range of problems in this field of research was studied and further explored as part of the Deutsche Forschungsgemeinschaft (DFG) priority program 1253 on “Optimization with Partial Differential Equations” from 2006 to 2013. The investigations were motivated by the fascinating potential applications and challenging mathematical problems that arise in the field of PDE constrained optimization. New analytic and algorithmic paradigms have been developed, implemented and validated in the context of real-world applications. In this special volume, contributions from more than fifteen German universities combine the results of this interdisciplinary program with a focus on applied mathematics. The book is divided into five sections on “Constrained Optimization, Identification and Control”...
Order-constrained linear optimization.
Tidwell, Joe W; Dougherty, Michael R; Chrabaszcz, Jeffrey S; Thomas, Rick P
2017-11-01
Despite the fact that data and theories in the social, behavioural, and health sciences are often represented on an ordinal scale, there has been relatively little emphasis on modelling ordinal properties. The most common analytic framework used in psychological science is the general linear model, whose variants include ANOVA, MANOVA, and ordinary linear regression. While these methods are designed to provide the best fit to the metric properties of the data, they are not designed to maximally model ordinal properties. In this paper, we develop an order-constrained linear least-squares (OCLO) optimization algorithm that maximizes the linear least-squares fit to the data conditional on maximizing the ordinal fit based on Kendall's τ. The algorithm builds on the maximum rank correlation estimator (Han, 1987, Journal of Econometrics, 35, 303) and the general monotone model (Dougherty & Thomas, 2012, Psychological Review, 119, 321). Analyses of simulated data indicate that when modelling data that adhere to the assumptions of ordinary least squares, OCLO shows minimal bias, little increase in variance, and almost no loss in out-of-sample predictive accuracy. In contrast, under conditions in which data include a small number of extreme scores (fat-tailed distributions), OCLO shows less bias and variance, and substantially better out-of-sample predictive accuracy, even when the outliers are removed. We show that the advantages of OCLO over ordinary least squares in predicting new observations hold across a variety of scenarios in which researchers must decide to retain or eliminate extreme scores when fitting data. © 2017 The British Psychological Society.
Security constrained optimal power flow by modern optimization tools
African Journals Online (AJOL)
Security constrained optimal power flow by modern optimization tools. ... International Journal of Engineering, Science and Technology ... If you would like more information about how to print, save, and work with PDFs, Highwire Press ...
Neuroevolutionary Constrained Optimization for Content Creation
DEFF Research Database (Denmark)
Liapis, Antonios; Yannakakis, Georgios N.; Togelius, Julian
2011-01-01
and thruster types and topologies) independently of game physics and steering strategies. According to the proposed framework, the designer picks a set of requirements for the spaceship that a constrained optimizer attempts to satisfy. The constraint satisfaction approach followed is based on neuroevolution...... and survival tasks and are also visually appealing....
Constrained Optimization and Optimal Control for Partial Differential Equations
Leugering, Günter; Griewank, Andreas
2012-01-01
This special volume focuses on optimization and control of processes governed by partial differential equations. The contributors are mostly participants of the DFG-priority program 1253: Optimization with PDE-constraints which is active since 2006. The book is organized in sections which cover almost the entire spectrum of modern research in this emerging field. Indeed, even though the field of optimal control and optimization for PDE-constrained problems has undergone a dramatic increase of interest during the last four decades, a full theory for nonlinear problems is still lacking. The cont
Bidirectional Dynamic Diversity Evolutionary Algorithm for Constrained Optimization
Directory of Open Access Journals (Sweden)
Weishang Gao
2013-01-01
Full Text Available Evolutionary algorithms (EAs were shown to be effective for complex constrained optimization problems. However, inflexible exploration-exploitation and improper penalty in EAs with penalty function would lead to losing the global optimum nearby or on the constrained boundary. To determine an appropriate penalty coefficient is also difficult in most studies. In this paper, we propose a bidirectional dynamic diversity evolutionary algorithm (Bi-DDEA with multiagents guiding exploration-exploitation through local extrema to the global optimum in suitable steps. In Bi-DDEA potential advantage is detected by three kinds of agents. The scale and the density of agents will change dynamically according to the emerging of potential optimal area, which play an important role of flexible exploration-exploitation. Meanwhile, a novel double optimum estimation strategy with objective fitness and penalty fitness is suggested to compute, respectively, the dominance trend of agents in feasible region and forbidden region. This bidirectional evolving with multiagents can not only effectively avoid the problem of determining penalty coefficient but also quickly converge to the global optimum nearby or on the constrained boundary. By examining the rapidity and veracity of Bi-DDEA across benchmark functions, the proposed method is shown to be effective.
Constrained optimization via simulation models for new product innovation
Pujowidianto, Nugroho A.
2017-11-01
We consider the problem of constrained optimization where the decision makers aim to optimize the primary performance measure while constraining the secondary performance measures. This paper provides a brief overview of stochastically constrained optimization via discrete event simulation. Most review papers tend to be methodology-based. This review attempts to be problem-based as decision makers may have already decided on the problem formulation. We consider constrained optimization models as there are usually constraints on secondary performance measures as trade-off in new product development. It starts by laying out different possible methods and the reasons using constrained optimization via simulation models. It is then followed by the review of different simulation optimization approach to address constrained optimization depending on the number of decision variables, the type of constraints, and the risk preferences of the decision makers in handling uncertainties.
Krohling, Renato A; Coelho, Leandro dos Santos
2006-12-01
In this correspondence, an approach based on coevolutionary particle swarm optimization to solve constrained optimization problems formulated as min-max problems is presented. In standard or canonical particle swarm optimization (PSO), a uniform probability distribution is used to generate random numbers for the accelerating coefficients of the local and global terms. We propose a Gaussian probability distribution to generate the accelerating coefficients of PSO. Two populations of PSO using Gaussian distribution are used on the optimization algorithm that is tested on a suite of well-known benchmark constrained optimization problems. Results have been compared with the canonical PSO (constriction factor) and with a coevolutionary genetic algorithm. Simulation results show the suitability of the proposed algorithm in terms of effectiveness and robustness.
On benchmarking Stochastic Global Optimization Algorithms
Hendrix, E.M.T.; Lancinskas, A.
2015-01-01
A multitude of heuristic stochastic optimization algorithms have been described in literature to obtain good solutions of the box-constrained global optimization problem often with a limit on the number of used function evaluations. In the larger question of which algorithms behave well on which
Stochastic and global optimization
National Research Council Canada - National Science Library
Dzemyda, Gintautas; Šaltenis, Vydūnas; Zhilinskas, A; Mockus, Jonas
2002-01-01
... and Effectiveness of Controlled Random Search E. M. T. Hendrix, P. M. Ortigosa and I. García 129 9. Discrete Backtracking Adaptive Search for Global Optimization B. P. Kristinsdottir, Z. B. Zabinsky and...
Optimization of an implicit constrained multi-physics system for motor wheels of electric vehicle
International Nuclear Information System (INIS)
Lei, Fei; Du, Bin; Liu, Xin; Xie, Xiaoping; Chai, Tian
2016-01-01
In this paper, implicit constrained multi-physics model of a motor wheel for an electric vehicle is built and then optimized. A novel optimization approach is proposed to solve the compliance problem between implicit constraints and stochastic global optimization. Firstly, multi-physics model of motor wheel is built from the theories of structural mechanics, electromagnetism and thermal physics. Then, implicit constraints are applied from the vehicle performances and magnetic characteristics. Implicit constrained optimization is carried out by a series of unconstrained optimization and verifications. In practice, sequentially updated subspaces are designed to completely substitute the original design space in local areas. In each subspace, a solution is obtained and is then verified by the implicit constraints. Optimal solutions which satisfy the implicit constraints are accepted as final candidates. The final global optimal solution is optimized from those candidates. Discussions are carried out to discover the differences between optimal solutions with unconstrained problem and different implicit constrained problems. Results show that the implicit constraints have significant influences on the optimal solution and the proposed approach is effective in finding the optimals. - Highlights: • An implicit constrained multi-physics model is built for sizing a motor wheel. • Vehicle dynamic performances are applied as implicit constraints for nonlinear system. • An efficient novel optimization is proposed to explore the constrained design space. • The motor wheel is optimized to achieve maximum efficiency on vehicle dynamics. • Influences of implicit constraints on vehicle performances are compared and analyzed.
A model for optimal constrained adaptive testing
van der Linden, Willem J.; Reese, Lynda M.
2001-01-01
A model for constrained computerized adaptive testing is proposed in which the information on the test at the ability estimate is maximized subject to a large variety of possible constraints on the contents of the test. At each item-selection step, a full test is first assembled to have maximum
A model for optimal constrained adaptive testing
van der Linden, Willem J.; Reese, Lynda M.
1997-01-01
A model for constrained computerized adaptive testing is proposed in which the information in the test at the ability estimate is maximized subject to a large variety of possible constraints on the contents of the test. At each item-selection step, a full test is first assembled to have maximum
Chance constrained uncertain classification via robust optimization
Ben-Tal, A.; Bhadra, S.; Bhattacharayya, C.; Saketha Nat, J.
2011-01-01
This paper studies the problem of constructing robust classifiers when the training is plagued with uncertainty. The problem is posed as a Chance-Constrained Program (CCP) which ensures that the uncertain data points are classified correctly with high probability. Unfortunately such a CCP turns out
Filter Pattern Search Algorithms for Mixed Variable Constrained Optimization Problems
National Research Council Canada - National Science Library
Abramson, Mark A; Audet, Charles; Dennis, Jr, J. E
2004-01-01
.... This class combines and extends the Audet-Dennis Generalized Pattern Search (GPS) algorithms for bound constrained mixed variable optimization, and their GPS-filter algorithms for general nonlinear constraints...
Constrained Dynamic Optimality and Binomial Terminal Wealth
DEFF Research Database (Denmark)
Pedersen, J. L.; Peskir, G.
2018-01-01
with interest rate $r \\in {R}$). Letting $P_{t,x}$ denote a probability measure under which $X^u$ takes value $x$ at time $t,$ we study the dynamic version of the nonlinear optimal control problem $\\inf_u\\, Var{t,X_t^u}(X_T^u)$ where the infimum is taken over admissible controls $u$ subject to $X_t^u \\ge e...... a martingale method combined with Lagrange multipliers, we derive the dynamically optimal control $u_*^d$ in closed form and prove that the dynamically optimal terminal wealth $X_T^d$ can only take two values $g$ and $\\beta$. This binomial nature of the dynamically optimal strategy stands in sharp contrast...... with other known portfolio selection strategies encountered in the literature. A direct comparison shows that the dynamically optimal (time-consistent) strategy outperforms the statically optimal (time-inconsistent) strategy in the problem....
Pareto-optimal estimates that constrain mean California precipitation change
Langenbrunner, B.; Neelin, J. D.
2017-12-01
Global climate model (GCM) projections of greenhouse gas-induced precipitation change can exhibit notable uncertainty at the regional scale, particularly in regions where the mean change is small compared to internal variability. This is especially true for California, which is located in a transition zone between robust precipitation increases to the north and decreases to the south, and where GCMs from the Climate Model Intercomparison Project phase 5 (CMIP5) archive show no consensus on mean change (in either magnitude or sign) across the central and southern parts of the state. With the goal of constraining this uncertainty, we apply a multiobjective approach to a large set of subensembles (subsets of models from the full CMIP5 ensemble). These constraints are based on subensemble performance in three fields important to California precipitation: tropical Pacific sea surface temperatures, upper-level zonal winds in the midlatitude Pacific, and precipitation over the state. An evolutionary algorithm is used to sort through and identify the set of Pareto-optimal subensembles across these three measures in the historical climatology, and we use this information to constrain end-of-century California wet season precipitation change. This technique narrows the range of projections throughout the state and increases confidence in estimates of positive mean change. Furthermore, these methods complement and generalize emergent constraint approaches that aim to restrict uncertainty in end-of-century projections, and they have applications to even broader aspects of uncertainty quantification, including parameter sensitivity and model calibration.
A one-layer recurrent neural network for constrained nonsmooth invex optimization.
Li, Guocheng; Yan, Zheng; Wang, Jun
2014-02-01
Invexity is an important notion in nonconvex optimization. In this paper, a one-layer recurrent neural network is proposed for solving constrained nonsmooth invex optimization problems, designed based on an exact penalty function method. It is proved herein that any state of the proposed neural network is globally convergent to the optimal solution set of constrained invex optimization problems, with a sufficiently large penalty parameter. In addition, any neural state is globally convergent to the unique optimal solution, provided that the objective function and constraint functions are pseudoconvex. Moreover, any neural state is globally convergent to the feasible region in finite time and stays there thereafter. The lower bounds of the penalty parameter and convergence time are also estimated. Two numerical examples are provided to illustrate the performances of the proposed neural network. Copyright © 2013 Elsevier Ltd. All rights reserved.
Constrained Optimization of MIMO Training Sequences
Directory of Open Access Journals (Sweden)
Coon Justin P
2007-01-01
Full Text Available Multiple-input multiple-output (MIMO systems have shown a huge potential for increased spectral efficiency and throughput. With an increasing number of transmitting antennas comes the burden of providing training for channel estimation for coherent detection. In some special cases optimal, in the sense of mean-squared error (MSE, training sequences have been designed. However, in many practical systems it is not feasible to analytically find optimal solutions and numerical techniques must be used. In this paper, two systems (unique word (UW single carrier and OFDM with nulled subcarriers are considered and a method of designing near-optimal training sequences using nonlinear optimization techniques is proposed. In particular, interior-point (IP algorithms such as the barrier method are discussed. Although the two systems seem unrelated, the cost function, which is the MSE of the channel estimate, is shown to be effectively the same for each scenario. Also, additional constraints, such as peak-to-average power ratio (PAPR, are considered and shown to be easily included in the optimization process. Numerical examples illustrate the effectiveness of the designed training sequences, both in terms of MSE and bit-error rate (BER.
Chance-constrained optimization of demand response to price signals
DEFF Research Database (Denmark)
Dorini, Gianluca Fabio; Pinson, Pierre; Madsen, Henrik
2013-01-01
within a recursive least squares (RLS) framework using data measurable at the grid level, in an adaptive fashion. Optimal price signals are generated by embedding the FIR models within a chance-constrained optimization framework. The objective is to keep the price signal as unchanged as possible from...
Effective Teaching of Economics: A Constrained Optimization Problem?
Hultberg, Patrik T.; Calonge, David Santandreu
2017-01-01
One of the fundamental tenets of economics is that decisions are often the result of optimization problems subject to resource constraints. Consumers optimize utility, subject to constraints imposed by prices and income. As economics faculty, instructors attempt to maximize student learning while being constrained by their own and students'…
Sequential unconstrained minimization algorithms for constrained optimization
International Nuclear Information System (INIS)
Byrne, Charles
2008-01-01
The problem of minimizing a function f(x):R J → R, subject to constraints on the vector variable x, occurs frequently in inverse problems. Even without constraints, finding a minimizer of f(x) may require iterative methods. We consider here a general class of iterative algorithms that find a solution to the constrained minimization problem as the limit of a sequence of vectors, each solving an unconstrained minimization problem. Our sequential unconstrained minimization algorithm (SUMMA) is an iterative procedure for constrained minimization. At the kth step we minimize the function G k (x)=f(x)+g k (x), to obtain x k . The auxiliary functions g k (x):D subset of R J → R + are nonnegative on the set D, each x k is assumed to lie within D, and the objective is to minimize the continuous function f:R J → R over x in the set C = D-bar, the closure of D. We assume that such minimizers exist, and denote one such by x-circumflex. We assume that the functions g k (x) satisfy the inequalities 0≤g k (x)≤G k-1 (x)-G k-1 (x k-1 ), for k = 2, 3, .... Using this assumption, we show that the sequence {(x k )} is decreasing and converges to f(x-circumflex). If the restriction of f(x) to D has bounded level sets, which happens if x-circumflex is unique and f(x) is closed, proper and convex, then the sequence {x k } is bounded, and f(x*)=f(x-circumflex), for any cluster point x*. Therefore, if x-circumflex is unique, x* = x-circumflex and {x k } → x-circumflex. When x-circumflex is not unique, convergence can still be obtained, in particular cases. The SUMMA includes, as particular cases, the well-known barrier- and penalty-function methods, the simultaneous multiplicative algebraic reconstruction technique (SMART), the proximal minimization algorithm of Censor and Zenios, the entropic proximal methods of Teboulle, as well as certain cases of gradient descent and the Newton–Raphson method. The proof techniques used for SUMMA can be extended to obtain related results
New Exact Penalty Functions for Nonlinear Constrained Optimization Problems
Directory of Open Access Journals (Sweden)
Bingzhuang Liu
2014-01-01
Full Text Available For two kinds of nonlinear constrained optimization problems, we propose two simple penalty functions, respectively, by augmenting the dimension of the primal problem with a variable that controls the weight of the penalty terms. Both of the penalty functions enjoy improved smoothness. Under mild conditions, it can be proved that our penalty functions are both exact in the sense that local minimizers of the associated penalty problem are precisely the local minimizers of the original constrained problem.
Constrained ripple optimization of Tokamak bundle divertors
International Nuclear Information System (INIS)
Hively, L.M.; Rome, J.A.; Lynch, V.E.; Lyon, J.F.; Fowler, R.H.; Peng, Y-K.M.; Dory, R.A.
1983-02-01
Magnetic field ripple from a tokamak bundle divertor is localized to a small toroidal sector and must be treated differently from the usual (distributed) toroidal field (TF) coil ripple. Generally, in a tokamak with an unoptimized divertor design, all of the banana-trapped fast ions are quickly lost due to banana drift diffusion or to trapping between the 1/R variation in absolute value vector B ω B and local field maxima due to the divertor. A computer code has been written to optimize automatically on-axis ripple subject to these constraints, while varying up to nine design parameters. Optimum configurations have low on-axis ripple ( 0 ) are lost. However, because finite-sized TF coils have not been used in this study, the flux bundle is not expanded
A Simply Constrained Optimization Reformulation of KKT Systems Arising from Variational Inequalities
International Nuclear Information System (INIS)
Facchinei, F.; Fischer, A.; Kanzow, C.; Peng, J.-M.
1999-01-01
The Karush-Kuhn-Tucker (KKT) conditions can be regarded as optimality conditions for both variational inequalities and constrained optimization problems. In order to overcome some drawbacks of recently proposed reformulations of KKT systems, we propose casting KKT systems as a minimization problem with nonnegativity constraints on some of the variables. We prove that, under fairly mild assumptions, every stationary point of this constrained minimization problem is a solution of the KKT conditions. Based on this reformulation, a new algorithm for the solution of the KKT conditions is suggested and shown to have some strong global and local convergence properties
Security constrained optimal power flow by modern optimization tools
African Journals Online (AJOL)
The main objective of an optimal power flow (OPF) functions is to optimize .... It is characterized as propagation of plants and this happens by gametes union. ... ss and different variables, for example, wind, nearby fertilization can have a critic.
Evolutionary global optimization, manifolds and applications
Aguiar e Oliveira Junior, Hime
2016-01-01
This book presents powerful techniques for solving global optimization problems on manifolds by means of evolutionary algorithms, and shows in practice how these techniques can be applied to solve real-world problems. It describes recent findings and well-known key facts in general and differential topology, revisiting them all in the context of application to current optimization problems. Special emphasis is put on game theory problems. Here, these problems are reformulated as constrained global optimization tasks and solved with the help of Fuzzy ASA. In addition, more abstract examples, including minimizations of well-known functions, are also included. Although the Fuzzy ASA approach has been chosen as the main optimizing paradigm, the book suggests that other metaheuristic methods could be used as well. Some of them are introduced, together with their advantages and disadvantages. Readers should possess some knowledge of linear algebra, and of basic concepts of numerical analysis and probability theory....
Optimal Power Constrained Distributed Detection over a Noisy Multiaccess Channel
Directory of Open Access Journals (Sweden)
Zhiwen Hu
2015-01-01
Full Text Available The problem of optimal power constrained distributed detection over a noisy multiaccess channel (MAC is addressed. Under local power constraints, we define the transformation function for sensor to realize the mapping from local decision to transmitted waveform. The deflection coefficient maximization (DCM is used to optimize the performance of power constrained fusion system. Using optimality conditions, we derive the closed-form solution to the considered problem. Monte Carlo simulations are carried out to evaluate the performance of the proposed new method. Simulation results show that the proposed method could significantly improve the detection performance of the fusion system with low signal-to-noise ratio (SNR. We also show that the proposed new method has a robust detection performance for broad SNR region.
Directory of Open Access Journals (Sweden)
Hailong Wang
2018-01-01
Full Text Available The backtracking search optimization algorithm (BSA is a population-based evolutionary algorithm for numerical optimization problems. BSA has a powerful global exploration capacity while its local exploitation capability is relatively poor. This affects the convergence speed of the algorithm. In this paper, we propose a modified BSA inspired by simulated annealing (BSAISA to overcome the deficiency of BSA. In the BSAISA, the amplitude control factor (F is modified based on the Metropolis criterion in simulated annealing. The redesigned F could be adaptively decreased as the number of iterations increases and it does not introduce extra parameters. A self-adaptive ε-constrained method is used to handle the strict constraints. We compared the performance of the proposed BSAISA with BSA and other well-known algorithms when solving thirteen constrained benchmarks and five engineering design problems. The simulation results demonstrated that BSAISA is more effective than BSA and more competitive with other well-known algorithms in terms of convergence speed.
Quasicanonical structure of optimal control in constrained discrete systems
Sieniutycz, S.
2003-06-01
This paper considers discrete processes governed by difference rather than differential equations for the state transformation. The basic question asked is if and when Hamiltonian canonical structures are possible in optimal discrete systems. Considering constrained discrete control, general optimization algorithms are derived that constitute suitable theoretical and computational tools when evaluating extremum properties of constrained physical models. The mathematical basis of the general theory is the Bellman method of dynamic programming (DP) and its extension in the form of the so-called Carathéodory-Boltyanski (CB) stage criterion which allows a variation of the terminal state that is otherwise fixed in the Bellman's method. Two relatively unknown, powerful optimization algorithms are obtained: an unconventional discrete formalism of optimization based on a Hamiltonian for multistage systems with unconstrained intervals of holdup time, and the time interval constrained extension of the formalism. These results are general; namely, one arrives at: the discrete canonical Hamilton equations, maximum principles, and (at the continuous limit of processes with free intervals of time) the classical Hamilton-Jacobi theory along with all basic results of variational calculus. Vast spectrum of applications of the theory is briefly discussed.
Superalloy design - A Monte Carlo constrained optimization method
CSIR Research Space (South Africa)
Stander, CM
1996-01-01
Full Text Available optimization method C. M. Stander Division of Materials Science and Technology, CSIR, PO Box 395, Pretoria, Republic of South Africa Received 74 March 1996; accepted 24 June 1996 A method, based on Monte Carlo constrained... successful hit, i.e. when Liow < LMP,,, < Lhiph, and for all the properties, Pj?, < P, < Pi@?. If successful this hit falls within the ROA. Repeat steps 4 and 5 to find at least ten (or more) successful hits with values...
Nonlinear Chance Constrained Problems: Optimality Conditions, Regularization and Solvers
Czech Academy of Sciences Publication Activity Database
Adam, Lukáš; Branda, Martin
2016-01-01
Roč. 170, č. 2 (2016), s. 419-436 ISSN 0022-3239 R&D Projects: GA ČR GA15-00735S Institutional support: RVO:67985556 Keywords : Chance constrained programming * Optimality conditions * Regularization * Algorithms * Free MATLAB codes Subject RIV: BB - Applied Statistics, Operational Research Impact factor: 1.289, year: 2016 http://library.utia.cas.cz/separaty/2016/MTR/adam-0460909.pdf
Constrained multi-objective optimization of storage ring lattices
Husain, Riyasat; Ghodke, A. D.
2018-03-01
The storage ring lattice optimization is a class of constrained multi-objective optimization problem, where in addition to low beam emittance, a large dynamic aperture for good injection efficiency and improved beam lifetime are also desirable. The convergence and computation times are of great concern for the optimization algorithms, as various objectives are to be optimized and a number of accelerator parameters to be varied over a large span with several constraints. In this paper, a study of storage ring lattice optimization using differential evolution is presented. The optimization results are compared with two most widely used optimization techniques in accelerators-genetic algorithm and particle swarm optimization. It is found that the differential evolution produces a better Pareto optimal front in reasonable computation time between two conflicting objectives-beam emittance and dispersion function in the straight section. The differential evolution was used, extensively, for the optimization of linear and nonlinear lattices of Indus-2 for exploring various operational modes within the magnet power supply capabilities.
Subspace Barzilai-Borwein Gradient Method for Large-Scale Bound Constrained Optimization
International Nuclear Information System (INIS)
Xiao Yunhai; Hu Qingjie
2008-01-01
An active set subspace Barzilai-Borwein gradient algorithm for large-scale bound constrained optimization is proposed. The active sets are estimated by an identification technique. The search direction consists of two parts: some of the components are simply defined; the other components are determined by the Barzilai-Borwein gradient method. In this work, a nonmonotone line search strategy that guarantees global convergence is used. Preliminary numerical results show that the proposed method is promising, and competitive with the well-known method SPG on a subset of bound constrained problems from CUTEr collection
Stress-constrained topology optimization for compliant mechanism design
DEFF Research Database (Denmark)
de Leon, Daniel M.; Alexandersen, Joe; Jun, Jun S.
2015-01-01
This article presents an application of stress-constrained topology optimization to compliant mechanism design. An output displacement maximization formulation is used, together with the SIMP approach and a projection method to ensure convergence to nearly discrete designs. The maximum stress...... is approximated using a normalized version of the commonly-used p-norm of the effective von Mises stresses. The usual problems associated with topology optimization for compliant mechanism design: one-node and/or intermediate density hinges are alleviated by the stress constraint. However, it is also shown...
Thermally-Constrained Fuel-Optimal ISS Maneuvers
Bhatt, Sagar; Svecz, Andrew; Alaniz, Abran; Jang, Jiann-Woei; Nguyen, Louis; Spanos, Pol
2015-01-01
Optimal Propellant Maneuvers (OPMs) are now being used to rotate the International Space Station (ISS) and have saved hundreds of kilograms of propellant over the last two years. The savings are achieved by commanding the ISS to follow a pre-planned attitude trajectory optimized to take advantage of environmental torques. The trajectory is obtained by solving an optimal control problem. Prior to use on orbit, OPM trajectories are screened to ensure a static sun vector (SSV) does not occur during the maneuver. The SSV is an indicator that the ISS hardware temperatures may exceed thermal limits, causing damage to the components. In this paper, thermally-constrained fuel-optimal trajectories are presented that avoid an SSV and can be used throughout the year while still reducing propellant consumption significantly.
Global optimization and simulated annealing
Dekkers, A.; Aarts, E.H.L.
1988-01-01
In this paper we are concerned with global optimization, which can be defined as the problem of finding points on a bounded subset of Rn in which some real valued functionf assumes its optimal (i.e. maximal or minimal) value. We present a stochastic approach which is based on the simulated annealing
Adaptive Multi-Agent Systems for Constrained Optimization
Macready, William; Bieniawski, Stefan; Wolpert, David H.
2004-01-01
Product Distribution (PD) theory is a new framework for analyzing and controlling distributed systems. Here we demonstrate its use for distributed stochastic optimization. First we review one motivation of PD theory, as the information-theoretic extension of conventional full-rationality game theory to the case of bounded rational agents. In this extension the equilibrium of the game is the optimizer of a Lagrangian of the (probability distribution of) the joint state of the agents. When the game in question is a team game with constraints, that equilibrium optimizes the expected value of the team game utility, subject to those constraints. The updating of the Lagrange parameters in the Lagrangian can be viewed as a form of automated annealing, that focuses the MAS more and more on the optimal pure strategy. This provides a simple way to map the solution of any constrained optimization problem onto the equilibrium of a Multi-Agent System (MAS). We present computer experiments involving both the Queen s problem and K-SAT validating the predictions of PD theory and its use for off-the-shelf distributed adaptive optimization.
Convex analysis and global optimization
Tuy, Hoang
2016-01-01
This book presents state-of-the-art results and methodologies in modern global optimization, and has been a staple reference for researchers, engineers, advanced students (also in applied mathematics), and practitioners in various fields of engineering. The second edition has been brought up to date and continues to develop a coherent and rigorous theory of deterministic global optimization, highlighting the essential role of convex analysis. The text has been revised and expanded to meet the needs of research, education, and applications for many years to come. Updates for this new edition include: · Discussion of modern approaches to minimax, fixed point, and equilibrium theorems, and to nonconvex optimization; · Increased focus on dealing more efficiently with ill-posed problems of global optimization, particularly those with hard constraints;
Acceleration techniques in the univariate Lipschitz global optimization
Sergeyev, Yaroslav D.; Kvasov, Dmitri E.; Mukhametzhanov, Marat S.; De Franco, Angela
2016-10-01
Univariate box-constrained Lipschitz global optimization problems are considered in this contribution. Geometric and information statistical approaches are presented. The novel powerful local tuning and local improvement techniques are described in the contribution as well as the traditional ways to estimate the Lipschitz constant. The advantages of the presented local tuning and local improvement techniques are demonstrated using the operational characteristics approach for comparing deterministic global optimization algorithms on the class of 100 widely used test functions.
Fast optimization of statistical potentials for structurally constrained phylogenetic models
Directory of Open Access Journals (Sweden)
Rodrigue Nicolas
2009-09-01
Full Text Available Abstract Background Statistical approaches for protein design are relevant in the field of molecular evolutionary studies. In recent years, new, so-called structurally constrained (SC models of protein-coding sequence evolution have been proposed, which use statistical potentials to assess sequence-structure compatibility. In a previous work, we defined a statistical framework for optimizing knowledge-based potentials especially suited to SC models. Our method used the maximum likelihood principle and provided what we call the joint potentials. However, the method required numerical estimations by the use of computationally heavy Markov Chain Monte Carlo sampling algorithms. Results Here, we develop an alternative optimization procedure, based on a leave-one-out argument coupled to fast gradient descent algorithms. We assess that the leave-one-out potential yields very similar results to the joint approach developed previously, both in terms of the resulting potential parameters, and by Bayes factor evaluation in a phylogenetic context. On the other hand, the leave-one-out approach results in a considerable computational benefit (up to a 1,000 fold decrease in computational time for the optimization procedure. Conclusion Due to its computational speed, the optimization method we propose offers an attractive alternative for the design and empirical evaluation of alternative forms of potentials, using large data sets and high-dimensional parameterizations.
Block-triangular preconditioners for PDE-constrained optimization
Rees, Tyrone; Stoll, Martin
2010-01-01
In this paper we investigate the possibility of using a block-triangular preconditioner for saddle point problems arising in PDE-constrained optimization. In particular, we focus on a conjugate gradient-type method introduced by Bramble and Pasciak that uses self-adjointness of the preconditioned system in a non-standard inner product. We show when the Chebyshev semi-iteration is used as a preconditioner for the relevant matrix blocks involving the finite element mass matrix that the main drawback of the Bramble-Pasciak method-the appropriate scaling of the preconditioners-is easily overcome. We present an eigenvalue analysis for the block-triangular preconditioners that gives convergence bounds in the non-standard inner product and illustrates their competitiveness on a number of computed examples. Copyright © 2010 John Wiley & Sons, Ltd.
Block-triangular preconditioners for PDE-constrained optimization
Rees, Tyrone
2010-11-26
In this paper we investigate the possibility of using a block-triangular preconditioner for saddle point problems arising in PDE-constrained optimization. In particular, we focus on a conjugate gradient-type method introduced by Bramble and Pasciak that uses self-adjointness of the preconditioned system in a non-standard inner product. We show when the Chebyshev semi-iteration is used as a preconditioner for the relevant matrix blocks involving the finite element mass matrix that the main drawback of the Bramble-Pasciak method-the appropriate scaling of the preconditioners-is easily overcome. We present an eigenvalue analysis for the block-triangular preconditioners that gives convergence bounds in the non-standard inner product and illustrates their competitiveness on a number of computed examples. Copyright © 2010 John Wiley & Sons, Ltd.
Optimal dispatch in dynamic security constrained open power market
International Nuclear Information System (INIS)
Singh, S.N.; David, A.K.
2002-01-01
Power system security is a new concern in the competitive power market operation, because the integration of the system controller and the generation owner has been broken. This paper presents an approach for dynamic security constrained optimal dispatch in restructured power market environment. The transient energy margin using transient energy function (TEF) approach has been used to calculate the stability margin of the system and a hybrid method is applied to calculate the approximate unstable equilibrium point (UEP) that is used to calculate the exact UEP and thus, the energy margin using TEF. The case study results illustrated on two systems shows that the operating mechanisms are compatible with the new business environment. (author)
A New Interpolation Approach for Linearly Constrained Convex Optimization
Espinoza, Francisco
2012-08-01
In this thesis we propose a new class of Linearly Constrained Convex Optimization methods based on the use of a generalization of Shepard\\'s interpolation formula. We prove the properties of the surface such as the interpolation property at the boundary of the feasible region and the convergence of the gradient to the null space of the constraints at the boundary. We explore several descent techniques such as steepest descent, two quasi-Newton methods and the Newton\\'s method. Moreover, we implement in the Matlab language several versions of the method, particularly for the case of Quadratic Programming with bounded variables. Finally, we carry out performance tests against Matab Optimization Toolbox methods for convex optimization and implementations of the standard log-barrier and active-set methods. We conclude that the steepest descent technique seems to be the best choice so far for our method and that it is competitive with other standard methods both in performance and empirical growth order.
Li, Wei; Ciais, Philippe; Wang, Yilong; Yin, Yi; Peng, Shushi; Zhu, Zaichun; Bastos, Ana; Yue, Chao; Ballantyne, Ashley P.; Broquet, Grégoire; Canadell, Josep G.; Cescatti, Alessandro; Chen, Chi; Cooper, Leila; Friedlingstein, Pierre; Le Quéré, Corinne; Myneni, Ranga B.; Piao, Shilong
2018-01-01
To assess global carbon cycle variability, we decompose the net land carbon sink into the sum of gross primary productivity (GPP), terrestrial ecosystem respiration (TER), and fire emissions and apply a Bayesian framework to constrain these fluxes between 1980 and 2014. The constrained GPP and TER fluxes show an increasing trend of only half of the prior trend simulated by models. From the optimization, we infer that TER increased in parallel with GPP from 1980 to 1990, but then stalled during the cooler periods, in 1990-1994 coincident with the Pinatubo eruption, and during the recent warming hiatus period. After each of these TER stalling periods, TER is found to increase faster than GPP, explaining a relative reduction of the net land sink. These results shed light on decadal variations of GPP and TER and suggest that they exhibit different responses to temperature anomalies over the last 35 years.
Constraining global methane emissions and uptake by ecosystems
International Nuclear Information System (INIS)
Spahni, R.; Wania, R.; Neef, L.; Van Weele, M.; Van Velthoven, P.; Pison, I.; Bousquet, P.
2011-01-01
Natural methane (CH 4 ) emissions from wet ecosystems are an important part of today's global CH 4 budget. Climate affects the exchange of CH 4 between ecosystems and the atmosphere by influencing CH 4 production, oxidation, and transport in the soil. The net CH 4 exchange depends on ecosystem hydrology, soil and vegetation characteristics. Here, the LPJ-WHyMe global dynamical vegetation model is used to simulate global net CH 4 emissions for different ecosystems: northern peat-lands (45 degrees-90 degrees N), naturally inundated wetlands (60 degrees S-45 degrees N), rice agriculture and wet mineral soils. Mineral soils are a potential CH 4 sink, but can also be a source with the direction of the net exchange depending on soil moisture content. The geographical and seasonal distributions are evaluated against multi-dimensional atmospheric inversions for 2003-2005, using two independent four-dimensional variational assimilation systems. The atmospheric inversions are constrained by the atmospheric CH 4 observations of the SCIAMACHY satellite instrument and global surface networks. Compared to LPJ-WHyMe the inversions result in a significant reduction in the emissions from northern peat-lands and suggest that LPJ-WHyMe maximum annual emissions peak about one month late. The inversions do not put strong constraints on the division of sources between inundated wetlands and wet mineral soils in the tropics. Based on the inversion results we diagnose model parameters in LPJ-WHyMe and simulate the surface exchange of CH 4 over the period 1990-2008. Over the whole period we infer an increase of global ecosystem CH 4 emissions of +1.11 TgCH 4 yr -1 , not considering potential additional changes in wetland extent. The increase in simulated CH 4 emissions is attributed to enhanced soil respiration resulting from the observed rise in land temperature and in atmospheric carbon dioxide that were used as input. The long term decline of the atmospheric CH 4 growth rate from 1990
Electrochemomechanical constrained multiobjective optimization of PPy/MWCNT actuators
International Nuclear Information System (INIS)
Khalili, N; Naguib, H E; Kwon, R H
2014-01-01
Polypyrrole (PPy) conducting polymers have shown a great potential for the fabrication of conjugated polymer-based actuating devices. Consequently, they have been a key point in developing many advanced emerging applications such as biomedical devices and biomimetic robotics. When designing an actuator, taking all of the related decision variables, their roles and relationships into consideration is of pivotal importance to determine the actuator’s final performance. Therefore, the central focus of this study is to develop an electrochemomechanical constrained multiobjective optimization model of a PPy/MWCNTs trilayer actuator. For this purpose, the objective functions are designed to capture the three main characteristics of these actuators, namely their tip vertical displacement, blocking force and response time. To obtain the optimum range of the designated decision variables within the feasible domain, a multiobjective optimization algorithm is applied while appropriate constraints are imposed. The optimum points form a Pareto surface on which they are consistently spread. The numerical results are presented; these results enable one to design an actuator with consideration to the desired output performances. For the experimental analysis, a multilayer bending-type actuator is fabricated, which is composed of a PVDF layer and two layers of PPy with an incorporated layer of multi-walled carbon nanotubes deposited on each side of the PVDF membrane. The numerical results are experimentally verified; in order to determine the performance of the fabricated actuator, its outputs are compared with a neat PPy actuator’s experimental and numerical counterparts. (paper)
Keulen, van T.A.C.; Gillot, J.; Jager, de A.G.; Steinbuch, M.
2014-01-01
This paper presents a numerical solution for scalar state constrained optimal control problems. The algorithm rewrites the constrained optimal control problem as a sequence of unconstrained optimal control problems which can be solved recursively as a two point boundary value problem. The solution
Stability Constrained Efficiency Optimization for Droop Controlled DC-DC Conversion System
DEFF Research Database (Denmark)
Meng, Lexuan; Dragicevic, Tomislav; Guerrero, Josep M.
2013-01-01
implementing tertiary regulation. Moreover, system dynamic is affected when shifting VRs. Therefore, the stability is considered in optimization by constraining the eigenvalues arising from dynamic state space model of the system. Genetic algorithm is used in searching for global efficiency optimum while....... As the efficiency of each converter changes with output power, virtual resistances (VRs) are set as decision variables for adjusting power sharing proportion among converters. It is noteworthy that apart from restoring the voltage deviation, secondary control plays an important role to stabilize dc bus voltage when...
Global optimization and sensitivity analysis
International Nuclear Information System (INIS)
Cacuci, D.G.
1990-01-01
A new direction for the analysis of nonlinear models of nuclear systems is suggested to overcome fundamental limitations of sensitivity analysis and optimization methods currently prevalent in nuclear engineering usage. This direction is toward a global analysis of the behavior of the respective system as its design parameters are allowed to vary over their respective design ranges. Presented is a methodology for global analysis that unifies and extends the current scopes of sensitivity analysis and optimization by identifying all the critical points (maxima, minima) and solution bifurcation points together with corresponding sensitivities at any design point of interest. The potential applicability of this methodology is illustrated with test problems involving multiple critical points and bifurcations and comprising both equality and inequality constraints
Vorozheikin, A.; Gonchar, T.; Panfilov, I.; Sopov, E.; Sopov, S.
2009-01-01
A new algorithm for the solution of complex constrained optimization problems based on the probabilistic genetic algorithm with optimal solution prediction is proposed. The efficiency investigation results in comparison with standard genetic algorithm are presented.
Peng, Guoyi; Cao, Shuliang; Ishizuka, Masaru; Hayama, Shinji
2002-06-01
This paper is concerned with the design optimization of axial flow hydraulic turbine runner blade geometry. In order to obtain a better design plan with good performance, a new comprehensive performance optimization procedure has been presented by combining a multi-variable multi-objective constrained optimization model with a Q3D inverse computation and a performance prediction procedure. With careful analysis of the inverse design of axial hydraulic turbine runner, the total hydraulic loss and the cavitation coefficient are taken as optimization objectives and a comprehensive objective function is defined using the weight factors. Parameters of a newly proposed blade bound circulation distribution function and parameters describing positions of blade leading and training edges in the meridional flow passage are taken as optimization variables.The optimization procedure has been applied to the design optimization of a Kaplan runner with specific speed of 440 kW. Numerical results show that the performance of designed runner is successfully improved through optimization computation. The optimization model is found to be validated and it has the feature of good convergence. With the multi-objective optimization model, it is possible to control the performance of designed runner by adjusting the value of weight factors defining the comprehensive objective function. Copyright
Depletion mapping and constrained optimization to support managing groundwater extraction
Fienen, Michael N.; Bradbury, Kenneth R.; Kniffin, Maribeth; Barlow, Paul M.
2018-01-01
Groundwater models often serve as management tools to evaluate competing water uses including ecosystems, irrigated agriculture, industry, municipal supply, and others. Depletion potential mapping—showing the model-calculated potential impacts that wells have on stream baseflow—can form the basis for multiple potential management approaches in an oversubscribed basin. Specific management approaches can include scenarios proposed by stakeholders, systematic changes in well pumping based on depletion potential, and formal constrained optimization, which can be used to quantify the tradeoff between water use and stream baseflow. Variables such as the maximum amount of reduction allowed in each well and various groupings of wells using, for example, K-means clustering considering spatial proximity and depletion potential are considered. These approaches provide a potential starting point and guidance for resource managers and stakeholders to make decisions about groundwater management in a basin, spreading responsibility in different ways. We illustrate these approaches in the Little Plover River basin in central Wisconsin, United States—home to a rich agricultural tradition, with farmland and urban areas both in close proximity to a groundwater-dependent trout stream. Groundwater withdrawals have reduced baseflow supplying the Little Plover River below a legally established minimum. The techniques in this work were developed in response to engaged stakeholders with various interests and goals for the basin. They sought to develop a collaborative management plan at a watershed scale that restores the flow rate in the river in a manner that incorporates principles of shared governance and results in effective and minimally disruptive changes in groundwater extraction practices.
Directory of Open Access Journals (Sweden)
Vivek Patel
2012-08-01
Full Text Available Nature inspired population based algorithms is a research field which simulates different natural phenomena to solve a wide range of problems. Researchers have proposed several algorithms considering different natural phenomena. Teaching-Learning-based optimization (TLBO is one of the recently proposed population based algorithm which simulates the teaching-learning process of the class room. This algorithm does not require any algorithm-specific control parameters. In this paper, elitism concept is introduced in the TLBO algorithm and its effect on the performance of the algorithm is investigated. The effects of common controlling parameters such as the population size and the number of generations on the performance of the algorithm are also investigated. The proposed algorithm is tested on 35 constrained benchmark functions with different characteristics and the performance of the algorithm is compared with that of other well known optimization algorithms. The proposed algorithm can be applied to various optimization problems of the industrial environment.
International Nuclear Information System (INIS)
Nguyen, Q H; Choi, S B
2012-01-01
This research focuses on optimal design of different types of magnetorheological brakes (MRBs), from which an optimal selection of MRB types is identified. In the optimization, common types of MRB such as disc-type, drum-type, hybrid-types, and T-shaped type are considered. The optimization problem is to find the optimal value of significant geometric dimensions of the MRB that can produce a maximum braking torque. The MRB is constrained in a cylindrical volume of a specific radius and length. After a brief description of the configuration of MRB types, the braking torques of the MRBs are derived based on the Herschel–Bulkley model of the MR fluid. The optimal design of MRBs constrained in a specific cylindrical volume is then analysed. The objective of the optimization is to maximize the braking torque while the torque ratio (the ratio of maximum braking torque and the zero-field friction torque) is constrained to be greater than a certain value. A finite element analysis integrated with an optimization tool is employed to obtain optimal solutions of the MRBs. Optimal solutions of MRBs constrained in different volumes are obtained based on the proposed optimization procedure. From the results, discussions on the optimal selection of MRB types depending on constrained volumes are given. (paper)
Integrated Multidisciplinary Constrained Optimization of Offshore Support Structures
International Nuclear Information System (INIS)
Haghi, Rad; Molenaar, David P; Ashuri, Turaj; Van der Valk, Paul L C
2014-01-01
In the current offshore wind turbine support structure design method, the tower and foundation, which form the support structure are designed separately by the turbine and foundation designer. This method yields a suboptimal design and it results in a heavy, overdesigned and expensive support structure. This paper presents an integrated multidisciplinary approach to design the tower and foundation simultaneously. Aerodynamics, hydrodynamics, structure and soil mechanics are the modeled disciplines to capture the full dynamic behavior of the foundation and tower under different environmental conditions. The objective function to be minimized is the mass of the support structure. The model includes various design constraints: local and global buckling, modal frequencies, and fatigue damage along different stations of the structure. To show the usefulness of the method, an existing SWT-3.6-107 offshore wind turbine where its tower and foundation are designed separately is used as a case study. The result of the integrated multidisciplinary design optimization shows 12.1% reduction in the mass of the support structure, while satisfying all the design constraints
How market environment may constrain global franchising in emerging markets
Baena Graciá, Verónica
2011-01-01
Although emerging markets are some of the fastest growing economies in the world and represent countries that are experiencing a substantial economic transformation, little is known about the factors influencing country selection for expansion in those markets. In an attempt to enhance the knowledge that managers and scholars have on franchising expansion, the present study examines how market conditions may constrain international diffusion of franchising in emerging markets. They are: i) ge...
Fuzzy chance constrained linear programming model for scrap charge optimization in steel production
DEFF Research Database (Denmark)
Rong, Aiying; Lahdelma, Risto
2008-01-01
the uncertainty based on fuzzy set theory and constrain the failure risk based on a possibility measure. Consequently, the scrap charge optimization problem is modeled as a fuzzy chance constrained linear programming problem. Since the constraints of the model mainly address the specification of the product...
Century long observation constrained global dynamic downscaling and hydrologic implication
Kim, H.; Yoshimura, K.; Chang, E.; Famiglietti, J. S.; Oki, T.
2012-12-01
It has been suggested that greenhouse gas induced warming climate causes the acceleration of large scale hydrologic cycles, and, indeed, many regions on the Earth have been suffered by hydrologic extremes getting more frequent. However, historical observations are not able to provide enough information in comprehensive manner to understand their long-term variability and/or global distributions. In this study, a century long high resolution global climate data is developed in order to break through existing limitations. 20th Century Reanalysis (20CR) which has relatively low spatial resolution (~2.0°) and longer term availability (140 years) is dynamically downscaled into global T248 (~0.5°) resolution using Experimental Climate Prediction Center (ECPC) Global Spectral Model (GSM) by spectral nudging data assimilation technique. Also, Global Precipitation Climatology Centre (GPCC) and Climate Research Unit (CRU) observational data are adopted to reduce model dependent uncertainty. Downscaled product successfully represents realistic geographical detail keeping low frequency signal in mean state and spatiotemporal variability, while previous bias correction method fails to reproduce high frequency variability. Newly developed data is used to investigate how long-term large scale terrestrial hydrologic cycles have been changed globally and how they have been interacted with various climate modes, such as El-Niño Southern Oscillation (ENSO) and Atlantic Multidecadal Oscillation (AMO). As a further application, it will be used to provide atmospheric boundary condition of multiple land surface models in the Global Soil Wetness Project Phase 3 (GSWP3).
How Will Copper Contamination Constrain Future Global Steel Recycling?
Daehn, Katrin; Cabrera Serrenho, Andre; Allwood, Julian Mark
2017-01-01
Copper in steel causes metallurgical problems, but is pervasive in end-of-life scrap and cannot currently be removed commercially once in the melt. Contamination can be managed to an extent by globally trading scrap for use in tolerant applications and dilution with primary iron sources. However, the viability of long-term strategies can only be evaluated with a complete characterization of copper in the global steel system and this is presented in this paper. The copper concentration of flow...
A Framework for Constrained Optimization Problems Based on a Modified Particle Swarm Optimization
Directory of Open Access Journals (Sweden)
Biwei Tang
2016-01-01
Full Text Available This paper develops a particle swarm optimization (PSO based framework for constrained optimization problems (COPs. Aiming at enhancing the performance of PSO, a modified PSO algorithm, named SASPSO 2011, is proposed by adding a newly developed self-adaptive strategy to the standard particle swarm optimization 2011 (SPSO 2011 algorithm. Since the convergence of PSO is of great importance and significantly influences the performance of PSO, this paper first theoretically investigates the convergence of SASPSO 2011. Then, a parameter selection principle guaranteeing the convergence of SASPSO 2011 is provided. Subsequently, a SASPSO 2011-based framework is established to solve COPs. Attempting to increase the diversity of solutions and decrease optimization difficulties, the adaptive relaxation method, which is combined with the feasibility-based rule, is applied to handle constraints of COPs and evaluate candidate solutions in the developed framework. Finally, the proposed method is verified through 4 benchmark test functions and 2 real-world engineering problems against six PSO variants and some well-known methods proposed in the literature. Simulation results confirm that the proposed method is highly competitive in terms of the solution quality and can be considered as a vital alternative to solve COPs.
A Variant of the Topkis-Veinott Method for Solving Inequality Constrained Optimization Problems
International Nuclear Information System (INIS)
Birge, J. R.; Qi, L.; Wei, Z.
2000-01-01
In this paper we give a variant of the Topkis-Veinott method for solving inequality constrained optimization problems. This method uses a linearly constrained positive semidefinite quadratic problem to generate a feasible descent direction at each iteration. Under mild assumptions, the algorithm is shown to be globally convergent in the sense that every accumulation point of the sequence generated by the algorithm is a Fritz-John point of the problem. We introduce a Fritz-John (FJ) function, an FJ1 strong second-order sufficiency condition (FJ1-SSOSC), and an FJ2 strong second-order sufficiency condition (FJ2-SSOSC), and then show, without any constraint qualification (CQ), that (i) if an FJ point z satisfies the FJ1-SSOSC, then there exists a neighborhood N(z) of z such that, for any FJ point y element of N(z) {z } , f 0 (y) ≠ f 0 (z) , where f 0 is the objective function of the problem; (ii) if an FJ point z satisfies the FJ2-SSOSC, then z is a strict local minimum of the problem. The result (i) implies that the entire iteration point sequence generated by the method converges to an FJ point. We also show that if the parameters are chosen large enough, a unit step length can be accepted by the proposed algorithm
Artificial bee colony algorithm for constrained possibilistic portfolio optimization problem
Chen, Wei
2015-07-01
In this paper, we discuss the portfolio optimization problem with real-world constraints under the assumption that the returns of risky assets are fuzzy numbers. A new possibilistic mean-semiabsolute deviation model is proposed, in which transaction costs, cardinality and quantity constraints are considered. Due to such constraints the proposed model becomes a mixed integer nonlinear programming problem and traditional optimization methods fail to find the optimal solution efficiently. Thus, a modified artificial bee colony (MABC) algorithm is developed to solve the corresponding optimization problem. Finally, a numerical example is given to illustrate the effectiveness of the proposed model and the corresponding algorithm.
A One-Layer Recurrent Neural Network for Constrained Complex-Variable Convex Optimization.
Qin, Sitian; Feng, Jiqiang; Song, Jiahui; Wen, Xingnan; Xu, Chen
2018-03-01
In this paper, based on calculus and penalty method, a one-layer recurrent neural network is proposed for solving constrained complex-variable convex optimization. It is proved that for any initial point from a given domain, the state of the proposed neural network reaches the feasible region in finite time and converges to an optimal solution of the constrained complex-variable convex optimization finally. In contrast to existing neural networks for complex-variable convex optimization, the proposed neural network has a lower model complexity and better convergence. Some numerical examples and application are presented to substantiate the effectiveness of the proposed neural network.
How Will Copper Contamination Constrain Future Global Steel Recycling?
Daehn, Katrin E; Cabrera Serrenho, André; Allwood, Julian M
2017-06-06
Copper in steel causes metallurgical problems, but is pervasive in end-of-life scrap and cannot currently be removed commercially once in the melt. Contamination can be managed to an extent by globally trading scrap for use in tolerant applications and dilution with primary iron sources. However, the viability of long-term strategies can only be evaluated with a complete characterization of copper in the global steel system and this is presented in this paper. The copper concentration of flows along the 2008 steel supply chain is estimated from a survey of literature data and compared with estimates of the maximum concentration that can be tolerated in steel products. Estimates of final steel demand and scrap supply by sector are taken from a global stock-saturation model to determine when the amount of copper in the steel cycle will exceed that which can be tolerated. Best estimates show that quantities of copper arising from conventional scrap preparation can be managed in the global steel system until 2050 assuming perfectly coordinated trade and extensive dilution, but this strategy will become increasingly impractical. Technical and policy interventions along the supply chain are presented to close product loops before this global constraint.
Constraining neutron guide optimizations with phase-space considerations
Energy Technology Data Exchange (ETDEWEB)
Bertelsen, Mads, E-mail: mads.bertelsen@gmail.com; Lefmann, Kim
2016-09-11
We introduce a method named the Minimalist Principle that serves to reduce the parameter space for neutron guide optimization when the required beam divergence is limited. The reduced parameter space will restrict the optimization to guides with a minimal neutron intake that are still theoretically able to deliver the maximal possible performance. The geometrical constraints are derived using phase-space propagation from moderator to guide and from guide to sample, while assuming that the optimized guides will achieve perfect transport of the limited neutron intake. Guide systems optimized using these constraints are shown to provide performance close to guides optimized without any constraints, however the divergence received at the sample is limited to the desired interval, even when the neutron transport is not limited by the supermirrors used in the guide. As the constraints strongly limit the parameter space for the optimizer, two control parameters are introduced that can be used to adjust the selected subspace, effectively balancing between maximizing neutron transport and avoiding background from unnecessary neutrons. One parameter is needed to describe the expected focusing abilities of the guide to be optimized, going from perfectly focusing to no correlation between position and velocity. The second parameter controls neutron intake into the guide, so that one can select exactly how aggressively the background should be limited. We show examples of guides optimized using these constraints which demonstrates the higher signal to noise than conventional optimizations. Furthermore the parameter controlling neutron intake is explored which shows that the simulated optimal neutron intake is close to the analytically predicted, when assuming that the guide is dominated by multiple scattering events.
Essays and surveys in global optimization
Audet, Charles; Savard, Giles
2005-01-01
Global optimization aims at solving the most general problems of deterministic mathematical programming. In addition, once the solutions are found, this methodology is also expected to prove their optimality. With these difficulties in mind, global optimization is becoming an increasingly powerful and important methodology. This book is the most recent examination of its mathematical capability, power, and wide ranging solutions to many fields in the applied sciences.
Introduction to Nonlinear and Global Optimization
Hendrix, E.M.T.; Tóth, B.
2010-01-01
This self-contained text provides a solid introduction to global and nonlinear optimization, providing students of mathematics and interdisciplinary sciences with a strong foundation in applied optimization techniques. The book offers a unique hands-on and critical approach to applied optimization
International Nuclear Information System (INIS)
Yang Chao; Meza, Juan C.; Wang Linwang
2006-01-01
A new direct constrained optimization algorithm for minimizing the Kohn-Sham (KS) total energy functional is presented in this paper. The key ingredients of this algorithm involve projecting the total energy functional into a sequence of subspaces of small dimensions and seeking the minimizer of total energy functional within each subspace. The minimizer of a subspace energy functional not only provides a search direction along which the KS total energy functional decreases but also gives an optimal 'step-length' to move along this search direction. Numerical examples are provided to demonstrate that this new direct constrained optimization algorithm can be more efficient than the self-consistent field (SCF) iteration
Constrained Burn Optimization for the International Space Station
Brown, Aaron J.; Jones, Brandon A.
2017-01-01
In long-term trajectory planning for the International Space Station (ISS), translational burns are currently targeted sequentially to meet the immediate trajectory constraints, rather than simultaneously to meet all constraints, do not employ gradient-based search techniques, and are not optimized for a minimum total deltav (v) solution. An analytic formulation of the constraint gradients is developed and used in an optimization solver to overcome these obstacles. Two trajectory examples are explored, highlighting the advantage of the proposed method over the current approach, as well as the potential v and propellant savings in the event of propellant shortages.
Stochastic global optimization as a filtering problem
International Nuclear Information System (INIS)
Stinis, Panos
2012-01-01
We present a reformulation of stochastic global optimization as a filtering problem. The motivation behind this reformulation comes from the fact that for many optimization problems we cannot evaluate exactly the objective function to be optimized. Similarly, we may not be able to evaluate exactly the functions involved in iterative optimization algorithms. For example, we may only have access to noisy measurements of the functions or statistical estimates provided through Monte Carlo sampling. This makes iterative optimization algorithms behave like stochastic maps. Naive global optimization amounts to evolving a collection of realizations of this stochastic map and picking the realization with the best properties. This motivates the use of filtering techniques to allow focusing on realizations that are more promising than others. In particular, we present a filtering reformulation of global optimization in terms of a special case of sequential importance sampling methods called particle filters. The increasing popularity of particle filters is based on the simplicity of their implementation and their flexibility. We utilize the flexibility of particle filters to construct a stochastic global optimization algorithm which can converge to the optimal solution appreciably faster than naive global optimization. Several examples of parametric exponential density estimation are provided to demonstrate the efficiency of the approach.
Constrained optimal motion planning for autonomous vehicles using PRONTO
Aguiar, A.P.; Bayer, F.A.; Hauser, J.; Häusler, A.J.; Notarstefano, G.; Pascoal, A.M.; Rucco, A.; Saccon, A.
2017-01-01
This chapter provides an overview of the authors’ efforts in vehicle trajectory exploration and motion planning based on PRONTO, a numerical method for solving optimal control problems developed over the last two decades. The chapter reviews the basics of PRONTO, providing the appropriate references
Human fetal growth is constrained below optimal for perinatal survival
Vasak, B.; Koenen, S. V.; Koster, M. P. H.; Hukkelhoven, C. W. P. M.; Franx, A.; Hanson, M. A.; Visser, GHA
ObjectiveThe use of fetal growth charts assumes that the optimal size at birth is at the 50(th) birth-weight centile, but interaction between maternal constraints on fetal growth and the risks associated with small and large fetal size at birth may indicate that this assumption is not valid for
Efficient relaxations for joint chance constrained AC optimal power flow
Energy Technology Data Exchange (ETDEWEB)
Baker, Kyri; Toomey, Bridget
2017-07-01
Evolving power systems with increasing levels of stochasticity call for a need to solve optimal power flow problems with large quantities of random variables. Weather forecasts, electricity prices, and shifting load patterns introduce higher levels of uncertainty and can yield optimization problems that are difficult to solve in an efficient manner. Solution methods for single chance constraints in optimal power flow problems have been considered in the literature, ensuring single constraints are satisfied with a prescribed probability; however, joint chance constraints, ensuring multiple constraints are simultaneously satisfied, have predominantly been solved via scenario-based approaches or by utilizing Boole's inequality as an upper bound. In this paper, joint chance constraints are used to solve an AC optimal power flow problem while preventing overvoltages in distribution grids under high penetrations of photovoltaic systems. A tighter version of Boole's inequality is derived and used to provide a new upper bound on the joint chance constraint, and simulation results are shown demonstrating the benefit of the proposed upper bound. The new framework allows for a less conservative and more computationally efficient solution to considering joint chance constraints, specifically regarding preventing overvoltages.
Improved Sensitivity Relations in State Constrained Optimal Control
International Nuclear Information System (INIS)
Bettiol, Piernicola; Frankowska, Hélène; Vinter, Richard B.
2015-01-01
Sensitivity relations in optimal control provide an interpretation of the costate trajectory and the Hamiltonian, evaluated along an optimal trajectory, in terms of gradients of the value function. While sensitivity relations are a straightforward consequence of standard transversality conditions for state constraint free optimal control problems formulated in terms of control-dependent differential equations with smooth data, their verification for problems with either pathwise state constraints, nonsmooth data, or for problems where the dynamic constraint takes the form of a differential inclusion, requires careful analysis. In this paper we establish validity of both ‘full’ and ‘partial’ sensitivity relations for an adjoint state of the maximum principle, for optimal control problems with pathwise state constraints, where the underlying control system is described by a differential inclusion. The partial sensitivity relation interprets the costate in terms of partial Clarke subgradients of the value function with respect to the state variable, while the full sensitivity relation interprets the couple, comprising the costate and Hamiltonian, as the Clarke subgradient of the value function with respect to both time and state variables. These relations are distinct because, for nonsmooth data, the partial Clarke subdifferential does not coincide with the projection of the (full) Clarke subdifferential on the relevant coordinate space. We show for the first time (even for problems without state constraints) that a costate trajectory can be chosen to satisfy the partial and full sensitivity relations simultaneously. The partial sensitivity relation in this paper is new for state constraint problems, while the full sensitivity relation improves on earlier results in the literature (for optimal control problems formulated in terms of Lipschitz continuous multifunctions), because a less restrictive inward pointing hypothesis is invoked in the proof, and because
Advances in stochastic and deterministic global optimization
Zhigljavsky, Anatoly; Žilinskas, Julius
2016-01-01
Current research results in stochastic and deterministic global optimization including single and multiple objectives are explored and presented in this book by leading specialists from various fields. Contributions include applications to multidimensional data visualization, regression, survey calibration, inventory management, timetabling, chemical engineering, energy systems, and competitive facility location. Graduate students, researchers, and scientists in computer science, numerical analysis, optimization, and applied mathematics will be fascinated by the theoretical, computational, and application-oriented aspects of stochastic and deterministic global optimization explored in this book. This volume is dedicated to the 70th birthday of Antanas Žilinskas who is a leading world expert in global optimization. Professor Žilinskas's research has concentrated on studying models for the objective function, the development and implementation of efficient algorithms for global optimization with single and mu...
Preconditioning for partial differential equation constrained optimization with control constraints
Stoll, Martin; Wathen, Andy
2011-01-01
Optimal control problems with partial differential equations play an important role in many applications. The inclusion of bound constraints for the control poses a significant additional challenge for optimization methods. In this paper, we propose preconditioners for the saddle point problems that arise when a primal-dual active set method is used. We also show for this method that the same saddle point system can be derived when the method is considered as a semismooth Newton method. In addition, the projected gradient method can be employed to solve optimization problems with simple bounds, and we discuss the efficient solution of the linear systems in question. In the case when an acceleration technique is employed for the projected gradient method, this again yields a semismooth Newton method that is equivalent to the primal-dual active set method. We also consider the Moreau-Yosida regularization method for control constraints and efficient preconditioners for this technique. Numerical results illustrate the competitiveness of these approaches. © 2011 John Wiley & Sons, Ltd.
Preconditioning for partial differential equation constrained optimization with control constraints
Stoll, Martin
2011-10-18
Optimal control problems with partial differential equations play an important role in many applications. The inclusion of bound constraints for the control poses a significant additional challenge for optimization methods. In this paper, we propose preconditioners for the saddle point problems that arise when a primal-dual active set method is used. We also show for this method that the same saddle point system can be derived when the method is considered as a semismooth Newton method. In addition, the projected gradient method can be employed to solve optimization problems with simple bounds, and we discuss the efficient solution of the linear systems in question. In the case when an acceleration technique is employed for the projected gradient method, this again yields a semismooth Newton method that is equivalent to the primal-dual active set method. We also consider the Moreau-Yosida regularization method for control constraints and efficient preconditioners for this technique. Numerical results illustrate the competitiveness of these approaches. © 2011 John Wiley & Sons, Ltd.
Kinetic Constrained Optimization of the Golf Swing Hub Path
Directory of Open Access Journals (Sweden)
Steven M. Nesbit
2014-12-01
Full Text Available This study details an optimization of the golf swing, where the hand path and club angular trajectories are manipulated. The optimization goal was to maximize club head velocity at impact within the interaction kinetic limitations (force, torque, work, and power of the golfer as determined through the analysis of a typical swing using a two-dimensional dynamic model. The study was applied to four subjects with diverse swing capabilities and styles. It was determined that it is possible for all subjects to increase their club head velocity at impact within their respective kinetic limitations through combined modifications to their respective hand path and club angular trajectories. The manner of the modifications, the degree of velocity improvement, the amount of kinetic reduction, and the associated kinetic limitation quantities were subject dependent. By artificially minimizing selected kinetic inputs within the optimization algorithm, it was possible to identify swing trajectory characteristics that indicated relative kinetic weaknesses of a subject. Practical implications are offered based upon the findings of the study.
Constrained Fuzzy Predictive Control Using Particle Swarm Optimization
Directory of Open Access Journals (Sweden)
Oussama Ait Sahed
2015-01-01
Full Text Available A fuzzy predictive controller using particle swarm optimization (PSO approach is proposed. The aim is to develop an efficient algorithm that is able to handle the relatively complex optimization problem with minimal computational time. This can be achieved using reduced population size and small number of iterations. In this algorithm, instead of using the uniform distribution as in the conventional PSO algorithm, the initial particles positions are distributed according to the normal distribution law, within the area around the best position. The radius limiting this area is adaptively changed according to the tracking error values. Moreover, the choice of the initial best position is based on prior knowledge about the search space landscape and the fact that in most practical applications the dynamic optimization problem changes are gradual. The efficiency of the proposed control algorithm is evaluated by considering the control of the model of a 4 × 4 Multi-Input Multi-Output industrial boiler. This model is characterized by being nonlinear with high interactions between its inputs and outputs, having a nonminimum phase behaviour, and containing instabilities and time delays. The obtained results are compared to those of the control algorithms based on the conventional PSO and the linear approach.
Directory of Open Access Journals (Sweden)
Zheng Ling
2011-01-01
Full Text Available Damping treatments have been extensively used as a powerful means to damp out structural resonant vibrations. Usually, damping materials are fully covered on the surface of plates. The drawbacks of this conventional treatment are also obvious due to an added mass and excess material consumption. Therefore, it is not always economical and effective from an optimization design view. In this paper, a topology optimization approach is presented to maximize the modal damping ratio of the plate with constrained layer damping treatment. The governing equation of motion of the plate is derived on the basis of energy approach. A finite element model to describe dynamic performances of the plate is developed and used along with an optimization algorithm in order to determine the optimal topologies of constrained layer damping layout on the plate. The damping of visco-elastic layer is modeled by the complex modulus formula. Considering the vibration and energy dissipation mode of the plate with constrained layer damping treatment, damping material density and volume factor are considered as design variable and constraint respectively. Meantime, the modal damping ratio of the plate is assigned as the objective function in the topology optimization approach. The sensitivity of modal damping ratio to design variable is further derived and Method of Moving Asymptote (MMA is adopted to search the optimized topologies of constrained layer damping layout on the plate. Numerical examples are used to demonstrate the effectiveness of the proposed topology optimization approach. The results show that vibration energy dissipation of the plates can be enhanced by the optimal constrained layer damping layout. This optimal technology can be further extended to vibration attenuation of sandwich cylindrical shells which constitute the major building block of many critical structures such as cabins of aircrafts, hulls of submarines and bodies of rockets and missiles as an
Kinetic constrained optimization of the golf swing hub path.
Nesbit, Steven M; McGinnis, Ryan S
2014-12-01
This study details an optimization of the golf swing, where the hand path and club angular trajectories are manipulated. The optimization goal was to maximize club head velocity at impact within the interaction kinetic limitations (force, torque, work, and power) of the golfer as determined through the analysis of a typical swing using a two-dimensional dynamic model. The study was applied to four subjects with diverse swing capabilities and styles. It was determined that it is possible for all subjects to increase their club head velocity at impact within their respective kinetic limitations through combined modifications to their respective hand path and club angular trajectories. The manner of the modifications, the degree of velocity improvement, the amount of kinetic reduction, and the associated kinetic limitation quantities were subject dependent. By artificially minimizing selected kinetic inputs within the optimization algorithm, it was possible to identify swing trajectory characteristics that indicated relative kinetic weaknesses of a subject. Practical implications are offered based upon the findings of the study. Key PointsThe hand path trajectory is an important characteristic of the golf swing and greatly affects club head velocity and golfer/club energy transfer.It is possible to increase the energy transfer from the golfer to the club by modifying the hand path and swing trajectories without increasing the kinetic output demands on the golfer.It is possible to identify relative kinetic output strengths and weakness of a golfer through assessment of the hand path and swing trajectories.Increasing any one of the kinetic outputs of the golfer can potentially increase the club head velocity at impact.The hand path trajectory has important influences over the club swing trajectory.
Constrained optimization of test intervals using a steady-state genetic algorithm
International Nuclear Information System (INIS)
Martorell, S.; Carlos, S.; Sanchez, A.; Serradell, V.
2000-01-01
There is a growing interest from both the regulatory authorities and the nuclear industry to stimulate the use of Probabilistic Risk Analysis (PRA) for risk-informed applications at Nuclear Power Plants (NPPs). Nowadays, special attention is being paid on analyzing plant-specific changes to Test Intervals (TIs) within the Technical Specifications (TSs) of NPPs and it seems to be a consensus on the need of making these requirements more risk-effective and less costly. Resource versus risk-control effectiveness principles formally enters in optimization problems. This paper presents an approach for using the PRA models in conducting the constrained optimization of TIs based on a steady-state genetic algorithm (SSGA) where the cost or the burden is to be minimized while the risk or performance is constrained to be at a given level, or vice versa. The paper encompasses first with the problem formulation, where the objective function and constraints that apply in the constrained optimization of TIs based on risk and cost models at system level are derived. Next, the foundation of the optimizer is given, which is derived by customizing a SSGA in order to allow optimizing TIs under constraints. Also, a case study is performed using this approach, which shows the benefits of adopting both PRA models and genetic algorithms, in particular for the constrained optimization of TIs, although it is also expected a great benefit of using this approach to solve other engineering optimization problems. However, care must be taken in using genetic algorithms in constrained optimization problems as it is concluded in this paper
On the efficiency of chaos optimization algorithms for global optimization
International Nuclear Information System (INIS)
Yang Dixiong; Li Gang; Cheng Gengdong
2007-01-01
Chaos optimization algorithms as a novel method of global optimization have attracted much attention, which were all based on Logistic map. However, we have noticed that the probability density function of the chaotic sequences derived from Logistic map is a Chebyshev-type one, which may affect the global searching capacity and computational efficiency of chaos optimization algorithms considerably. Considering the statistical property of the chaotic sequences of Logistic map and Kent map, the improved hybrid chaos-BFGS optimization algorithm and the Kent map based hybrid chaos-BFGS algorithm are proposed. Five typical nonlinear functions with multimodal characteristic are tested to compare the performance of five hybrid optimization algorithms, which are the conventional Logistic map based chaos-BFGS algorithm, improved Logistic map based chaos-BFGS algorithm, Kent map based chaos-BFGS algorithm, Monte Carlo-BFGS algorithm, mesh-BFGS algorithm. The computational performance of the five algorithms is compared, and the numerical results make us question the high efficiency of the chaos optimization algorithms claimed in some references. It is concluded that the efficiency of the hybrid optimization algorithms is influenced by the statistical property of chaotic/stochastic sequences generated from chaotic/stochastic algorithms, and the location of the global optimum of nonlinear functions. In addition, it is inappropriate to advocate the high efficiency of the global optimization algorithms only depending on several numerical examples of low-dimensional functions
Directory of Open Access Journals (Sweden)
Minggang Dong
2014-01-01
Full Text Available Motivated by recent advancements in differential evolution and constraints handling methods, this paper presents a novel modified oracle penalty function-based composite differential evolution (MOCoDE for constrained optimization problems (COPs. More specifically, the original oracle penalty function approach is modified so as to satisfy the optimization criterion of COPs; then the modified oracle penalty function is incorporated in composite DE. Furthermore, in order to solve more complex COPs with discrete, integer, or binary variables, a discrete variable handling technique is introduced into MOCoDE to solve complex COPs with mix variables. This method is assessed on eleven constrained optimization benchmark functions and seven well-studied engineering problems in real life. Experimental results demonstrate that MOCoDE achieves competitive performance with respect to some other state-of-the-art approaches in constrained optimization evolutionary algorithms. Moreover, the strengths of the proposed method include few parameters and its ease of implementation, rendering it applicable to real life. Therefore, MOCoDE can be an efficient alternative to solving constrained optimization problems.
Shao, H.; Huang, Y.; Kolditz, O.
2015-12-01
Multiphase flow problems are numerically difficult to solve, as it often contains nonlinear Phase transition phenomena A conventional technique is to introduce the complementarity constraints where fluid properties such as liquid saturations are confined within a physically reasonable range. Based on such constraints, the mathematical model can be reformulated into a system of nonlinear partial differential equations coupled with variational inequalities. They can be then numerically handled by optimization algorithms. In this work, two different approaches utilizing the complementarity constraints based on persistent primary variables formulation[4] are implemented and investigated. The first approach proposed by Marchand et.al[1] is using "local complementary constraints", i.e. coupling the constraints with the local constitutive equations. The second approach[2],[3] , namely the "global complementary constrains", applies the constraints globally with the mass conservation equation. We will discuss how these two approaches are applied to solve non-isothermal componential multiphase flow problem with the phase change phenomenon. Several benchmarks will be presented for investigating the overall numerical performance of different approaches. The advantages and disadvantages of different models will also be concluded. References[1] E.Marchand, T.Mueller and P.Knabner. Fully coupled generalized hybrid-mixed finite element approximation of two-phase two-component flow in porous media. Part I: formulation and properties of the mathematical model, Computational Geosciences 17(2): 431-442, (2013). [2] A. Lauser, C. Hager, R. Helmig, B. Wohlmuth. A new approach for phase transitions in miscible multi-phase flow in porous media. Water Resour., 34,(2011), 957-966. [3] J. Jaffré, and A. Sboui. Henry's Law and Gas Phase Disappearance. Transp. Porous Media. 82, (2010), 521-526. [4] A. Bourgeat, M. Jurak and F. Smaï. Two-phase partially miscible flow and transport modeling in
Optimal beneficiation of global resources
Energy Technology Data Exchange (ETDEWEB)
Aloisi de Larderel, J. (Industry and Environment Office, Paris (France). United Nations Environment Programme)
1989-01-01
The growth of the world's population and related human activities are clearly leaving major effects on the environment and on the level of use of natural resources: forests are disappearing, air pollution is leading to acid rains, changes are occuring in the atmospheric ozone and global climate, more and more people lack access to reasonable safe supplies of water, soil pollution is becoming a problem, mineral and energy resources are increasingly being used. Producing more with less, producing more, polluting less, these are basic challenges that the world now faces. Low- and non-waste technologies are certainly one of the keys to those challenges.
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Farouq, S.; Neytcheva, M.
2017-01-01
Roč. 74, č. 1 (2017), s. 19-37 ISSN 1017-1398 Institutional support: RVO:68145535 Keywords : PDE-constrained optimization problems * finite elements * iterative solution method s * preconditioning Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.241, year: 2016 https://link.springer.com/article/10.1007%2Fs11075-016-0136-5
Mixed Integer PDE Constrained Optimization for the Control of a Wildfire Hazard
2017-01-01
Constrained Optimization for the Control of a Wildfire Hazard Herausgegeben von der Professor fur Angewandte Mathematik Professor Dr. rer. nat. Armin...and H.H. Tan . Finite difference methods for solving the two-dimensional advection-diffusion equation. Int. J. Numer. Meth. Fluids, 9:75-98, 1989. 6
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Farouq, S.; Neytcheva, M.
2017-01-01
Roč. 74, č. 1 (2017), s. 19-37 ISSN 1017-1398 Institutional support: RVO:68145535 Keywords : PDE-constrained optimization problems * finite elements * iterative solution methods * preconditioning Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.241, year: 2016 https://link.springer.com/article/10.1007%2Fs11075-016-0136-5
Metal artifact reduction in x-ray computed tomography (CT) by constrained optimization
International Nuclear Information System (INIS)
Zhang Xiaomeng; Wang Jing; Xing Lei
2011-01-01
Purpose: The streak artifacts caused by metal implants have long been recognized as a problem that limits various applications of CT imaging. In this work, the authors propose an iterative metal artifact reduction algorithm based on constrained optimization. Methods: After the shape and location of metal objects in the image domain is determined automatically by the binary metal identification algorithm and the segmentation of ''metal shadows'' in projection domain is done, constrained optimization is used for image reconstruction. It minimizes a predefined function that reflects a priori knowledge of the image, subject to the constraint that the estimated projection data are within a specified tolerance of the available metal-shadow-excluded projection data, with image non-negativity enforced. The minimization problem is solved through the alternation of projection-onto-convex-sets and the steepest gradient descent of the objective function. The constrained optimization algorithm is evaluated with a penalized smoothness objective. Results: The study shows that the proposed method is capable of significantly reducing metal artifacts, suppressing noise, and improving soft-tissue visibility. It outperforms the FBP-type methods and ART and EM methods and yields artifacts-free images. Conclusions: Constrained optimization is an effective way to deal with CT reconstruction with embedded metal objects. Although the method is presented in the context of metal artifacts, it is applicable to general ''missing data'' image reconstruction problems.
Volume-constrained optimization of magnetorheological and electrorheological valves and dampers
Rosenfeld, Nicholas C.; Wereley, Norman M.
2004-12-01
This paper presents a case study of magnetorheological (MR) and electrorheological (ER) valve design within a constrained cylindrical volume. The primary purpose of this study is to establish general design guidelines for volume-constrained MR valves. Additionally, this study compares the performance of volume-constrained MR valves against similarly constrained ER valves. Starting from basic design guidelines for an MR valve, a method for constructing candidate volume-constrained valve geometries is presented. A magnetic FEM program is then used to evaluate the magnetic properties of the candidate valves. An optimized MR valve is chosen by evaluating non-dimensional parameters describing the candidate valves' damping performance. A derivation of the non-dimensional damping coefficient for valves with both active and passive volumes is presented to allow comparison of valves with differing proportions of active and passive volumes. The performance of the optimized MR valve is then compared to that of a geometrically similar ER valve using both analytical and numerical techniques. An analytical equation relating the damping performances of geometrically similar MR and ER valves in as a function of fluid yield stresses and relative active fluid volume, and numerical calculations are provided to calculate each valve's damping performance and to validate the analytical calculations.
A Globally Convergent Matrix-Free Method for Constrained Equations and Its Linear Convergence Rate
Directory of Open Access Journals (Sweden)
Min Sun
2014-01-01
Full Text Available A matrix-free method for constrained equations is proposed, which is a combination of the well-known PRP (Polak-Ribière-Polyak conjugate gradient method and the famous hyperplane projection method. The new method is not only derivative-free, but also completely matrix-free, and consequently, it can be applied to solve large-scale constrained equations. We obtain global convergence of the new method without any differentiability requirement on the constrained equations. Compared with the existing gradient methods for solving such problem, the new method possesses linear convergence rate under standard conditions, and a relax factor γ is attached in the update step to accelerate convergence. Preliminary numerical results show that it is promising in practice.
Image denoising: Learning the noise model via nonsmooth PDE-constrained optimization
Reyes, Juan Carlos De los; Schö nlieb, Carola-Bibiane
2013-01-01
We propose a nonsmooth PDE-constrained optimization approach for the determination of the correct noise model in total variation (TV) image denoising. An optimization problem for the determination of the weights corresponding to different types of noise distributions is stated and existence of an optimal solution is proved. A tailored regularization approach for the approximation of the optimal parameter values is proposed thereafter and its consistency studied. Additionally, the differentiability of the solution operator is proved and an optimality system characterizing the optimal solutions of each regularized problem is derived. The optimal parameter values are numerically computed by using a quasi-Newton method, together with semismooth Newton type algorithms for the solution of the TV-subproblems. © 2013 American Institute of Mathematical Sciences.
Image denoising: Learning the noise model via nonsmooth PDE-constrained optimization
Reyes, Juan Carlos De los
2013-11-01
We propose a nonsmooth PDE-constrained optimization approach for the determination of the correct noise model in total variation (TV) image denoising. An optimization problem for the determination of the weights corresponding to different types of noise distributions is stated and existence of an optimal solution is proved. A tailored regularization approach for the approximation of the optimal parameter values is proposed thereafter and its consistency studied. Additionally, the differentiability of the solution operator is proved and an optimality system characterizing the optimal solutions of each regularized problem is derived. The optimal parameter values are numerically computed by using a quasi-Newton method, together with semismooth Newton type algorithms for the solution of the TV-subproblems. © 2013 American Institute of Mathematical Sciences.
Global optimization framework for solar building design
Silva, N.; Alves, N.; Pascoal-Faria, P.
2017-07-01
The generative modeling paradigm is a shift from static models to flexible models. It describes a modeling process using functions, methods and operators. The result is an algorithmic description of the construction process. Each evaluation of such an algorithm creates a model instance, which depends on its input parameters (width, height, volume, roof angle, orientation, location). These values are normally chosen according to aesthetic aspects and style. In this study, the model's parameters are automatically generated according to an objective function. A generative model can be optimized according to its parameters, in this way, the best solution for a constrained problem is determined. Besides the establishment of an overall framework design, this work consists on the identification of different building shapes and their main parameters, the creation of an algorithmic description for these main shapes and the formulation of the objective function, respecting a building's energy consumption (solar energy, heating and insulation). Additionally, the conception of an optimization pipeline, combining an energy calculation tool with a geometric scripting engine is presented. The methods developed leads to an automated and optimized 3D shape generation for the projected building (based on the desired conditions and according to specific constrains). The approach proposed will help in the construction of real buildings that account for less energy consumption and for a more sustainable world.
Liu, Qingshan; Wang, Jun
2011-04-01
This paper presents a one-layer recurrent neural network for solving a class of constrained nonsmooth optimization problems with piecewise-linear objective functions. The proposed neural network is guaranteed to be globally convergent in finite time to the optimal solutions under a mild condition on a derived lower bound of a single gain parameter in the model. The number of neurons in the neural network is the same as the number of decision variables of the optimization problem. Compared with existing neural networks for optimization, the proposed neural network has a couple of salient features such as finite-time convergence and a low model complexity. Specific models for two important special cases, namely, linear programming and nonsmooth optimization, are also presented. In addition, applications to the shortest path problem and constrained least absolute deviation problem are discussed with simulation results to demonstrate the effectiveness and characteristics of the proposed neural network.
A penalty method for PDE-constrained optimization in inverse problems
International Nuclear Information System (INIS)
Leeuwen, T van; Herrmann, F J
2016-01-01
Many inverse and parameter estimation problems can be written as PDE-constrained optimization problems. The goal is to infer the parameters, typically coefficients of the PDE, from partial measurements of the solutions of the PDE for several right-hand sides. Such PDE-constrained problems can be solved by finding a stationary point of the Lagrangian, which entails simultaneously updating the parameters and the (adjoint) state variables. For large-scale problems, such an all-at-once approach is not feasible as it requires storing all the state variables. In this case one usually resorts to a reduced approach where the constraints are explicitly eliminated (at each iteration) by solving the PDEs. These two approaches, and variations thereof, are the main workhorses for solving PDE-constrained optimization problems arising from inverse problems. In this paper, we present an alternative method that aims to combine the advantages of both approaches. Our method is based on a quadratic penalty formulation of the constrained optimization problem. By eliminating the state variable, we develop an efficient algorithm that has roughly the same computational complexity as the conventional reduced approach while exploiting a larger search space. Numerical results show that this method indeed reduces some of the nonlinearity of the problem and is less sensitive to the initial iterate. (paper)
Directory of Open Access Journals (Sweden)
R. Venkata Rao
2014-01-01
Full Text Available The present work proposes a multi-objective improved teaching-learning based optimization (MO-ITLBO algorithm for unconstrained and constrained multi-objective function optimization. The MO-ITLBO algorithm is the improved version of basic teaching-learning based optimization (TLBO algorithm adapted for multi-objective problems. The basic TLBO algorithm is improved to enhance its exploration and exploitation capacities by introducing the concept of number of teachers, adaptive teaching factor, tutorial training and self-motivated learning. The MO-ITLBO algorithm uses a grid-based approach to adaptively assess the non-dominated solutions (i.e. Pareto front maintained in an external archive. The performance of the MO-ITLBO algorithm is assessed by implementing it on unconstrained and constrained test problems proposed for the Congress on Evolutionary Computation 2009 (CEC 2009 competition. The performance assessment is done by using the inverted generational distance (IGD measure. The IGD measures obtained by using the MO-ITLBO algorithm are compared with the IGD measures of the other state-of-the-art algorithms available in the literature. Finally, Lexicographic ordering is used to assess the overall performance of competitive algorithms. Results have shown that the proposed MO-ITLBO algorithm has obtained the 1st rank in the optimization of unconstrained test functions and the 3rd rank in the optimization of constrained test functions.
A Direct Search Algorithm for Global Optimization
Directory of Open Access Journals (Sweden)
Enrique Baeyens
2016-06-01
Full Text Available A direct search algorithm is proposed for minimizing an arbitrary real valued function. The algorithm uses a new function transformation and three simplex-based operations. The function transformation provides global exploration features, while the simplex-based operations guarantees the termination of the algorithm and provides global convergence to a stationary point if the cost function is differentiable and its gradient is Lipschitz continuous. The algorithm’s performance has been extensively tested using benchmark functions and compared to some well-known global optimization algorithms. The results of the computational study show that the algorithm combines both simplicity and efficiency and is competitive with the heuristics-based strategies presently used for global optimization.
An L∞/L1-Constrained Quadratic Optimization Problem with Applications to Neural Networks
International Nuclear Information System (INIS)
Leizarowitz, Arie; Rubinstein, Jacob
2003-01-01
Pattern formation in associative neural networks is related to a quadratic optimization problem. Biological considerations imply that the functional is constrained in the L ∞ norm and in the L 1 norm. We consider such optimization problems. We derive the Euler-Lagrange equations, and construct basic properties of the maximizers. We study in some detail the case where the kernel of the quadratic functional is finite-dimensional. In this case the optimization problem can be fully characterized by the geometry of a certain convex and compact finite-dimensional set
Global optimization methods for engineering design
Arora, Jasbir S.
1990-01-01
The problem is to find a global minimum for the Problem P. Necessary and sufficient conditions are available for local optimality. However, global solution can be assured only under the assumption of convexity of the problem. If the constraint set S is compact and the cost function is continuous on it, existence of a global minimum is guaranteed. However, in view of the fact that no global optimality conditions are available, a global solution can be found only by an exhaustive search to satisfy Inequality. The exhaustive search can be organized in such a way that the entire design space need not be searched for the solution. This way the computational burden is reduced somewhat. It is concluded that zooming algorithm for global optimizations appears to be a good alternative to stochastic methods. More testing is needed; a general, robust, and efficient local minimizer is required. IDESIGN was used in all numerical calculations which is based on a sequential quadratic programming algorithm, and since feasible set keeps on shrinking, a good algorithm to find an initial feasible point is required. Such algorithms need to be developed and evaluated.
Conference on Convex Analysis and Global Optimization
Pardalos, Panos
2001-01-01
There has been much recent progress in global optimization algo rithms for nonconvex continuous and discrete problems from both a theoretical and a practical perspective. Convex analysis plays a fun damental role in the analysis and development of global optimization algorithms. This is due essentially to the fact that virtually all noncon vex optimization problems can be described using differences of convex functions and differences of convex sets. A conference on Convex Analysis and Global Optimization was held during June 5 -9, 2000 at Pythagorion, Samos, Greece. The conference was honoring the memory of C. Caratheodory (1873-1950) and was en dorsed by the Mathematical Programming Society (MPS) and by the Society for Industrial and Applied Mathematics (SIAM) Activity Group in Optimization. The conference was sponsored by the European Union (through the EPEAEK program), the Department of Mathematics of the Aegean University and the Center for Applied Optimization of the University of Florida, by th...
Liu, Qingshan; Guo, Zhishan; Wang, Jun
2012-02-01
In this paper, a one-layer recurrent neural network is proposed for solving pseudoconvex optimization problems subject to linear equality and bound constraints. Compared with the existing neural networks for optimization (e.g., the projection neural networks), the proposed neural network is capable of solving more general pseudoconvex optimization problems with equality and bound constraints. Moreover, it is capable of solving constrained fractional programming problems as a special case. The convergence of the state variables of the proposed neural network to achieve solution optimality is guaranteed as long as the designed parameters in the model are larger than the derived lower bounds. Numerical examples with simulation results illustrate the effectiveness and characteristics of the proposed neural network. In addition, an application for dynamic portfolio optimization is discussed. Copyright © 2011 Elsevier Ltd. All rights reserved.
Wang, Mingming; Luo, Jianjun; Yuan, Jianping; Walter, Ulrich
2018-05-01
Application of the multi-arm space robot will be more effective than single arm especially when the target is tumbling. This paper investigates the application of particle swarm optimization (PSO) strategy to coordinated trajectory planning of the dual-arm space robot in free-floating mode. In order to overcome the dynamics singularities issue, the direct kinematics equations in conjunction with constrained PSO are employed for coordinated trajectory planning of dual-arm space robot. The joint trajectories are parametrized with Bézier curve to simplify the calculation. Constrained PSO scheme with adaptive inertia weight is implemented to find the optimal solution of joint trajectories while specific objectives and imposed constraints are satisfied. The proposed method is not sensitive to the singularity issue due to the application of forward kinematic equations. Simulation results are presented for coordinated trajectory planning of two kinematically redundant manipulators mounted on a free-floating spacecraft and demonstrate the effectiveness of the proposed method.
On meeting capital requirements with a chance-constrained optimization model.
Atta Mills, Ebenezer Fiifi Emire; Yu, Bo; Gu, Lanlan
2016-01-01
This paper deals with a capital to risk asset ratio chance-constrained optimization model in the presence of loans, treasury bill, fixed assets and non-interest earning assets. To model the dynamics of loans, we introduce a modified CreditMetrics approach. This leads to development of a deterministic convex counterpart of capital to risk asset ratio chance constraint. We pursue the scope of analyzing our model under the worst-case scenario i.e. loan default. The theoretical model is analyzed by applying numerical procedures, in order to administer valuable insights from a financial outlook. Our results suggest that, our capital to risk asset ratio chance-constrained optimization model guarantees banks of meeting capital requirements of Basel III with a likelihood of 95 % irrespective of changes in future market value of assets.
A first-order multigrid method for bound-constrained convex optimization
Czech Academy of Sciences Publication Activity Database
Kočvara, Michal; Mohammed, S.
2016-01-01
Roč. 31, č. 3 (2016), s. 622-644 ISSN 1055-6788 R&D Projects: GA ČR(CZ) GAP201/12/0671 Grant - others:European Commission - EC(XE) 313781 Institutional support: RVO:67985556 Keywords : bound-constrained optimization * multigrid methods * linear complementarity problems Subject RIV: BA - General Mathematics Impact factor: 1.023, year: 2016 http://library.utia.cas.cz/separaty/2016/MTR/kocvara-0460326.pdf
He, Longfei; Xu, Zhaoguang; Niu, Zhanwen
2014-01-01
We focus on the joint production planning of complex supply chains facing stochastic demands and being constrained by carbon emission reduction policies. We pick two typical carbon emission reduction policies to research how emission regulation influences the profit and carbon footprint of a typical supply chain. We use the input-output model to capture the interrelated demand link between an arbitrary pair of two nodes in scenarios without or with carbon emission constraints. We design optim...
Directory of Open Access Journals (Sweden)
Kiran Teeparthi
2017-04-01
Full Text Available In this paper, a new low level with teamwork heterogeneous hybrid particle swarm optimization and artificial physics optimization (HPSO-APO algorithm is proposed to solve the multi-objective security constrained optimal power flow (MO-SCOPF problem. Being engaged with the environmental and total production cost concerns, wind energy is highly penetrating to the main grid. The total production cost, active power losses and security index are considered as the objective functions. These are simultaneously optimized using the proposed algorithm for base case and contingency cases. Though PSO algorithm exhibits good convergence characteristic, fails to give near optimal solution. On the other hand, the APO algorithm shows the capability of improving diversity in search space and also to reach a near global optimum point, whereas, APO is prone to premature convergence. The proposed hybrid HPSO-APO algorithm combines both individual algorithm strengths, to get balance between global and local search capability. The APO algorithm is improving diversity in the search space of the PSO algorithm. The hybrid optimization algorithm is employed to alleviate the line overloads by generator rescheduling during contingencies. The standard IEEE 30-bus and Indian 75-bus practical test systems are considered to evaluate the robustness of the proposed method. The simulation results reveal that the proposed HPSO-APO method is more efficient and robust than the standard PSO and APO methods in terms of getting diverse Pareto optimal solutions. Hence, the proposed hybrid method can be used for the large interconnected power system to solve MO-SCOPF problem with integration of wind and thermal generators.
A Novel Particle Swarm Optimization Algorithm for Global Optimization.
Wang, Chun-Feng; Liu, Kui
2016-01-01
Particle Swarm Optimization (PSO) is a recently developed optimization method, which has attracted interest of researchers in various areas due to its simplicity and effectiveness, and many variants have been proposed. In this paper, a novel Particle Swarm Optimization algorithm is presented, in which the information of the best neighbor of each particle and the best particle of the entire population in the current iteration is considered. Meanwhile, to avoid premature, an abandoned mechanism is used. Furthermore, for improving the global convergence speed of our algorithm, a chaotic search is adopted in the best solution of the current iteration. To verify the performance of our algorithm, standard test functions have been employed. The experimental results show that the algorithm is much more robust and efficient than some existing Particle Swarm Optimization algorithms.
Competing intelligent search agents in global optimization
Energy Technology Data Exchange (ETDEWEB)
Streltsov, S.; Vakili, P. [Boston Univ., MA (United States); Muchnik, I. [Rutgers Univ., Piscataway, NJ (United States)
1996-12-31
In this paper we present a new search methodology that we view as a development of intelligent agent approach to the analysis of complex system. The main idea is to consider search process as a competition mechanism between concurrent adaptive intelligent agents. Agents cooperate in achieving a common search goal and at the same time compete with each other for computational resources. We propose a statistical selection approach to resource allocation between agents that leads to simple and efficient on average index allocation policies. We use global optimization as the most general setting that encompasses many types of search problems, and show how proposed selection policies can be used to improve and combine various global optimization methods.
Global Optimization Ensemble Model for Classification Methods
Anwar, Hina; Qamar, Usman; Muzaffar Qureshi, Abdul Wahab
2014-01-01
Supervised learning is the process of data mining for deducing rules from training datasets. A broad array of supervised learning algorithms exists, every one of them with its own advantages and drawbacks. There are some basic issues that affect the accuracy of classifier while solving a supervised learning problem, like bias-variance tradeoff, dimensionality of input space, and noise in the input data space. All these problems affect the accuracy of classifier and are the reason that there is no global optimal method for classification. There is not any generalized improvement method that can increase the accuracy of any classifier while addressing all the problems stated above. This paper proposes a global optimization ensemble model for classification methods (GMC) that can improve the overall accuracy for supervised learning problems. The experimental results on various public datasets showed that the proposed model improved the accuracy of the classification models from 1% to 30% depending upon the algorithm complexity. PMID:24883382
Global Optimization Ensemble Model for Classification Methods
Directory of Open Access Journals (Sweden)
Hina Anwar
2014-01-01
Full Text Available Supervised learning is the process of data mining for deducing rules from training datasets. A broad array of supervised learning algorithms exists, every one of them with its own advantages and drawbacks. There are some basic issues that affect the accuracy of classifier while solving a supervised learning problem, like bias-variance tradeoff, dimensionality of input space, and noise in the input data space. All these problems affect the accuracy of classifier and are the reason that there is no global optimal method for classification. There is not any generalized improvement method that can increase the accuracy of any classifier while addressing all the problems stated above. This paper proposes a global optimization ensemble model for classification methods (GMC that can improve the overall accuracy for supervised learning problems. The experimental results on various public datasets showed that the proposed model improved the accuracy of the classification models from 1% to 30% depending upon the algorithm complexity.
Directory of Open Access Journals (Sweden)
R. Manam
2017-12-01
Full Text Available In this paper, a sensitive constrained integer linear programming approach is formulated for the optimal allocation of Phasor Measurement Units (PMUs in a power system network to obtain state estimation. In this approach, sensitive buses along with zero injection buses (ZIB are considered for optimal allocation of PMUs in the network to generate state estimation solutions. Sensitive buses are evolved from the mean of bus voltages subjected to increase of load consistently up to 50%. Sensitive buses are ranked in order to place PMUs. Sensitive constrained optimal PMU allocation in case of single line and no line contingency are considered in observability analysis to ensure protection and control of power system from abnormal conditions. Modeling of ZIB constraints is included to minimize the number of PMU network allocations. This paper presents optimal allocation of PMU at sensitive buses with zero injection modeling, considering cost criteria and redundancy to increase the accuracy of state estimation solution without losing observability of the whole system. Simulations are carried out on IEEE 14, 30 and 57 bus systems and results obtained are compared with traditional and other state estimation methods available in the literature, to demonstrate the effectiveness of the proposed method.
CLFs-based optimization control for a class of constrained visual servoing systems.
Song, Xiulan; Miaomiao, Fu
2017-03-01
In this paper, we use the control Lyapunov function (CLF) technique to present an optimized visual servo control method for constrained eye-in-hand robot visual servoing systems. With the knowledge of camera intrinsic parameters and depth of target changes, visual servo control laws (i.e. translation speed) with adjustable parameters are derived by image point features and some known CLF of the visual servoing system. The Fibonacci method is employed to online compute the optimal value of those adjustable parameters, which yields an optimized control law to satisfy constraints of the visual servoing system. The Lyapunov's theorem and the properties of CLF are used to establish stability of the constrained visual servoing system in the closed-loop with the optimized control law. One merit of the presented method is that there is no requirement of online calculating the pseudo-inverse of the image Jacobian's matrix and the homography matrix. Simulation and experimental results illustrated the effectiveness of the method proposed here. Copyright © 2016 ISA. Published by Elsevier Ltd. All rights reserved.
Solving global optimization problems on GPU cluster
Energy Technology Data Exchange (ETDEWEB)
Barkalov, Konstantin; Gergel, Victor; Lebedev, Ilya [Lobachevsky State University of Nizhni Novgorod, Gagarin Avenue 23, 603950 Nizhni Novgorod (Russian Federation)
2016-06-08
The paper contains the results of investigation of a parallel global optimization algorithm combined with a dimension reduction scheme. This allows solving multidimensional problems by means of reducing to data-independent subproblems with smaller dimension solved in parallel. The new element implemented in the research consists in using several graphic accelerators at different computing nodes. The paper also includes results of solving problems of well-known multiextremal test class GKLS on Lobachevsky supercomputer using tens of thousands of GPU cores.
DEFF Research Database (Denmark)
Wang, Yong; Cai, Zixing; Zhou, Yuren
2009-01-01
A novel approach to deal with numerical and engineering constrained optimization problems, which incorporates a hybrid evolutionary algorithm and an adaptive constraint-handling technique, is presented in this paper. The hybrid evolutionary algorithm simultaneously uses simplex crossover and two...... mutation operators to generate the offspring population. Additionally, the adaptive constraint-handling technique consists of three main situations. In detail, at each situation, one constraint-handling mechanism is designed based on current population state. Experiments on 13 benchmark test functions...... and four well-known constrained design problems verify the effectiveness and efficiency of the proposed method. The experimental results show that integrating the hybrid evolutionary algorithm with the adaptive constraint-handling technique is beneficial, and the proposed method achieves competitive...
A First-order Prediction-Correction Algorithm for Time-varying (Constrained) Optimization: Preprint
Energy Technology Data Exchange (ETDEWEB)
Dall-Anese, Emiliano [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Simonetto, Andrea [Universite catholique de Louvain
2017-07-25
This paper focuses on the design of online algorithms based on prediction-correction steps to track the optimal solution of a time-varying constrained problem. Existing prediction-correction methods have been shown to work well for unconstrained convex problems and for settings where obtaining the inverse of the Hessian of the cost function can be computationally affordable. The prediction-correction algorithm proposed in this paper addresses the limitations of existing methods by tackling constrained problems and by designing a first-order prediction step that relies on the Hessian of the cost function (and do not require the computation of its inverse). Analytical results are established to quantify the tracking error. Numerical simulations corroborate the analytical results and showcase performance and benefits of the algorithms.
Implementation and verification of global optimization benchmark problems
Posypkin, Mikhail; Usov, Alexander
2017-12-01
The paper considers the implementation and verification of a test suite containing 150 benchmarks for global deterministic box-constrained optimization. A C++ library for describing standard mathematical expressions was developed for this purpose. The library automate the process of generating the value of a function and its' gradient at a given point and the interval estimates of a function and its' gradient on a given box using a single description. Based on this functionality, we have developed a collection of tests for an automatic verification of the proposed benchmarks. The verification has shown that literary sources contain mistakes in the benchmarks description. The library and the test suite are available for download and can be used freely.
Implementation and verification of global optimization benchmark problems
Directory of Open Access Journals (Sweden)
Posypkin Mikhail
2017-12-01
Full Text Available The paper considers the implementation and verification of a test suite containing 150 benchmarks for global deterministic box-constrained optimization. A C++ library for describing standard mathematical expressions was developed for this purpose. The library automate the process of generating the value of a function and its’ gradient at a given point and the interval estimates of a function and its’ gradient on a given box using a single description. Based on this functionality, we have developed a collection of tests for an automatic verification of the proposed benchmarks. The verification has shown that literary sources contain mistakes in the benchmarks description. The library and the test suite are available for download and can be used freely.
International Nuclear Information System (INIS)
Azadani, E. Nasr; Hosseinian, S.H.; Moradzadeh, B.
2010-01-01
Competitive bidding for energy and ancillary services is increasingly recognized as an important part of electricity markets. In addition, the transmission capacity limits should be considered to optimize the total market cost. In this paper, a new approach based on constrained particle swarm optimization (CPSO) is developed to deal with the multi-product (energy and reserve) and multi-area electricity market dispatch problem. Constraint handling is based on particle ranking and uniform distribution. CPSO method offers a new solution for optimizing the total market cost in a multi-area competitive electricity market considering the system constraints. The proposed technique shows promising performance for smooth and non smooth cost function as well. Three different systems are examined to demonstrate the effectiveness and the accuracy of the proposed algorithm. (author)
Robust and Reliable Portfolio Optimization Formulation of a Chance Constrained Problem
Directory of Open Access Journals (Sweden)
Sengupta Raghu Nandan
2017-02-01
Full Text Available We solve a linear chance constrained portfolio optimization problem using Robust Optimization (RO method wherein financial script/asset loss return distributions are considered as extreme valued. The objective function is a convex combination of portfolio’s CVaR and expected value of loss return, subject to a set of randomly perturbed chance constraints with specified probability values. The robust deterministic counterpart of the model takes the form of Second Order Cone Programming (SOCP problem. Results from extensive simulation runs show the efficacy of our proposed models, as it helps the investor to (i utilize extensive simulation studies to draw insights into the effect of randomness in portfolio decision making process, (ii incorporate different risk appetite scenarios to find the optimal solutions for the financial portfolio allocation problem and (iii compare the risk and return profiles of the investments made in both deterministic as well as in uncertain and highly volatile financial markets.
Solving Multi-Resource Constrained Project Scheduling Problem using Ant Colony Optimization
Directory of Open Access Journals (Sweden)
Hsiang-Hsi Huang
2015-01-01
Full Text Available This paper applied Ant Colony Optimization (ACO to develop a resource constraints scheduling model to achieve the resource allocation optimization and the shortest completion time of a project under resource constraints and the activities precedence requirement for projects. Resource leveling is also discussed and has to be achieved under the resource allocation optimization in this research. Testing cases and examples adopted from the international test bank were studied for verifying the effectiveness of the proposed model. The results showed that the solutions of different cases all have a better performance within a reasonable time. These can be obtained through ACO algorithm under the same constrained conditions. A program was written for the proposed model that is able to automatically produce the project resource requirement figure after the project duration is solved.
Directory of Open Access Journals (Sweden)
Liyang Wang
2017-01-01
Full Text Available The application of biped robots is always trapped by their high energy consumption. This paper makes a contribution by optimizing the joint torques to decrease the energy consumption without changing the biped gaits. In this work, a constrained quadratic programming (QP problem for energy optimization is formulated. A neurodynamics-based solver is presented to solve the QP problem. Differing from the existing literatures, the proposed neurodynamics-based energy optimization (NEO strategy minimizes the energy consumption and guarantees the following three important constraints simultaneously: (i the force-moment equilibrium equation of biped robots, (ii frictions applied by each leg on the ground to hold the biped robot without slippage and tipping over, and (iii physical limits of the motors. Simulations demonstrate that the proposed strategy is effective for energy-efficient biped walking.
Directory of Open Access Journals (Sweden)
Zhanpeng Fang
2015-01-01
Full Text Available A topology optimization method is proposed to minimize the resonant response of plates with constrained layer damping (CLD treatment under specified broadband harmonic excitations. The topology optimization problem is formulated and the square of displacement resonant response in frequency domain at the specified point is considered as the objective function. Two sensitivity analysis methods are investigated and discussed. The derivative of modal damp ratio is not considered in the conventional sensitivity analysis method. An improved sensitivity analysis method considering the derivative of modal damp ratio is developed to improve the computational accuracy of the sensitivity. The evolutionary structural optimization (ESO method is used to search the optimal layout of CLD material on plates. Numerical examples and experimental results show that the optimal layout of CLD treatment on the plate from the proposed topology optimization using the conventional sensitivity analysis or the improved sensitivity analysis can reduce the displacement resonant response. However, the optimization method using the improved sensitivity analysis can produce a higher modal damping ratio than that using the conventional sensitivity analysis and develop a smaller displacement resonant response.
Fuzzy Constrained Predictive Optimal Control of High Speed Train with Actuator Dynamics
Directory of Open Access Journals (Sweden)
Xi Wang
2016-01-01
Full Text Available We investigate the problem of fuzzy constrained predictive optimal control of high speed train considering the effect of actuator dynamics. The dynamics feature of the high speed train is modeled as a cascade of cars connected by flexible couplers, and the formulation is mathematically transformed into a Takagi-Sugeno (T-S fuzzy model. The goal of this study is to design a state feedback control law at each decision step to enhance safety, comfort, and energy efficiency of high speed train subject to safety constraints on the control input. Based on Lyapunov stability theory, the problem of optimizing an upper bound on the cruise control cost function subject to input constraints is reduced to a convex optimization problem involving linear matrix inequalities (LMIs. Furthermore, we analyze the influences of second-order actuator dynamics on the fuzzy constrained predictive controller, which shows risk of potentially deteriorating the overall system. Employing backstepping method, an actuator compensator is proposed to accommodate for the influence of the actuator dynamics. The experimental results show that with the proposed approach high speed train can track the desired speed, the relative coupler displacement between the neighbouring cars is stable at the equilibrium state, and the influence of actuator dynamics is reduced, which demonstrate the validity and effectiveness of the proposed approaches.
International Nuclear Information System (INIS)
Bahadormanesh, Nikrouz; Rahat, Shayan; Yarali, Milad
2017-01-01
Highlights: • A multi-objective optimization for radial expander in Organic Rankine Cycles is implemented. • By using firefly algorithm, Pareto front based on the size of turbine and thermal efficiency is produced. • Tension and vibration constrains have a significant effect on optimum design points. - Abstract: Organic Rankine Cycles are viable energy conversion systems in sustainable energy systems due to their compatibility with low-temperature heat sources. In the present study, one dimensional model of radial expanders in conjunction with a thermodynamic model of organic Rankine cycles is prepared. After verification, by defining thermal efficiency of the cycle and size parameter of a radial turbine as the objective functions, a multi-objective optimization was conducted regarding tension and vibration constraints for 4 different organic working fluids (R22, R245fa, R236fa and N-Pentane). In addition to mass flow rate, evaporator temperature, maximum pressure of cycle and turbo-machinery design parameters are selected as the decision variables. Regarding Pareto fronts, by a little increase in size of radial expanders, it is feasible to reach high efficiency. Moreover, by assessing the distribution of decision variables, the variables that play a major role in trending between the objective functions are found. Effects of mechanical and vibration constrains on optimum decision variables are investigated. The results of optimization can be considered as an initial values for design of radial turbines for Organic Rankine Cycles.
PAPR-Constrained Pareto-Optimal Waveform Design for OFDM-STAP Radar
Energy Technology Data Exchange (ETDEWEB)
Sen, Satyabrata [ORNL
2014-01-01
We propose a peak-to-average power ratio (PAPR) constrained Pareto-optimal waveform design approach for an orthogonal frequency division multiplexing (OFDM) radar signal to detect a target using the space-time adaptive processing (STAP) technique. The use of an OFDM signal does not only increase the frequency diversity of our system, but also enables us to adaptively design the OFDM coefficients in order to further improve the system performance. First, we develop a parametric OFDM-STAP measurement model by considering the effects of signaldependent clutter and colored noise. Then, we observe that the resulting STAP-performance can be improved by maximizing the output signal-to-interference-plus-noise ratio (SINR) with respect to the signal parameters. However, in practical scenarios, the computation of output SINR depends on the estimated values of the spatial and temporal frequencies and target scattering responses. Therefore, we formulate a PAPR-constrained multi-objective optimization (MOO) problem to design the OFDM spectral parameters by simultaneously optimizing four objective functions: maximizing the output SINR, minimizing two separate Cramer-Rao bounds (CRBs) on the normalized spatial and temporal frequencies, and minimizing the trace of CRB matrix on the target scattering coefficients estimations. We present several numerical examples to demonstrate the achieved performance improvement due to the adaptive waveform design.
2014-01-01
Portfolio optimization (selection) problem is an important and hard optimization problem that, with the addition of necessary realistic constraints, becomes computationally intractable. Nature-inspired metaheuristics are appropriate for solving such problems; however, literature review shows that there are very few applications of nature-inspired metaheuristics to portfolio optimization problem. This is especially true for swarm intelligence algorithms which represent the newer branch of nature-inspired algorithms. No application of any swarm intelligence metaheuristics to cardinality constrained mean-variance (CCMV) portfolio problem with entropy constraint was found in the literature. This paper introduces modified firefly algorithm (FA) for the CCMV portfolio model with entropy constraint. Firefly algorithm is one of the latest, very successful swarm intelligence algorithm; however, it exhibits some deficiencies when applied to constrained problems. To overcome lack of exploration power during early iterations, we modified the algorithm and tested it on standard portfolio benchmark data sets used in the literature. Our proposed modified firefly algorithm proved to be better than other state-of-the-art algorithms, while introduction of entropy diversity constraint further improved results. PMID:24991645
Bacanin, Nebojsa; Tuba, Milan
2014-01-01
Portfolio optimization (selection) problem is an important and hard optimization problem that, with the addition of necessary realistic constraints, becomes computationally intractable. Nature-inspired metaheuristics are appropriate for solving such problems; however, literature review shows that there are very few applications of nature-inspired metaheuristics to portfolio optimization problem. This is especially true for swarm intelligence algorithms which represent the newer branch of nature-inspired algorithms. No application of any swarm intelligence metaheuristics to cardinality constrained mean-variance (CCMV) portfolio problem with entropy constraint was found in the literature. This paper introduces modified firefly algorithm (FA) for the CCMV portfolio model with entropy constraint. Firefly algorithm is one of the latest, very successful swarm intelligence algorithm; however, it exhibits some deficiencies when applied to constrained problems. To overcome lack of exploration power during early iterations, we modified the algorithm and tested it on standard portfolio benchmark data sets used in the literature. Our proposed modified firefly algorithm proved to be better than other state-of-the-art algorithms, while introduction of entropy diversity constraint further improved results.
A New Continuous-Time Equality-Constrained Optimization to Avoid Singularity.
Quan, Quan; Cai, Kai-Yuan
2016-02-01
In equality-constrained optimization, a standard regularity assumption is often associated with feasible point methods, namely, that the gradients of constraints are linearly independent. In practice, the regularity assumption may be violated. In order to avoid such a singularity, a new projection matrix is proposed based on which a feasible point method to continuous-time, equality-constrained optimization is developed. First, the equality constraint is transformed into a continuous-time dynamical system with solutions that always satisfy the equality constraint. Second, a new projection matrix without singularity is proposed to realize the transformation. An update (or say a controller) is subsequently designed to decrease the objective function along the solutions of the transformed continuous-time dynamical system. The invariance principle is then applied to analyze the behavior of the solution. Furthermore, the proposed method is modified to address cases in which solutions do not satisfy the equality constraint. Finally, the proposed optimization approach is applied to three examples to demonstrate its effectiveness.
OPTIMIZED PARTICLE SWARM OPTIMIZATION BASED DEADLINE CONSTRAINED TASK SCHEDULING IN HYBRID CLOUD
Directory of Open Access Journals (Sweden)
Dhananjay Kumar
2016-01-01
Full Text Available Cloud Computing is a dominant way of sharing of computing resources that can be configured and provisioned easily. Task scheduling in Hybrid cloud is a challenge as it suffers from producing the best QoS (Quality of Service when there is a high demand. In this paper a new resource allocation algorithm, to find the best External Cloud provider when the intermediate provider’s resources aren’t enough to satisfy the customer’s demand is proposed. The proposed algorithm called Optimized Particle Swarm Optimization (OPSO combines the two metaheuristic algorithms namely Particle Swarm Optimization and Ant Colony Optimization (ACO. These metaheuristic algorithms are used for the purpose of optimization in the search space of the required solution, to find the best resource from the pool of resources and to obtain maximum profit even when the number of tasks submitted for execution is very high. This optimization is performed to allocate job requests to internal and external cloud providers to obtain maximum profit. It helps to improve the system performance by improving the CPU utilization, and handle multiple requests at the same time. The simulation result shows that an OPSO yields 0.1% - 5% profit to the intermediate cloud provider compared with standard PSO and ACO algorithms and it also increases the CPU utilization by 0.1%.
A perturbed martingale approach to global optimization
Energy Technology Data Exchange (ETDEWEB)
Sarkar, Saikat [Computational Mechanics Lab, Department of Civil Engineering, Indian Institute of Science, Bangalore 560012 (India); Roy, Debasish, E-mail: royd@civil.iisc.ernet.in [Computational Mechanics Lab, Department of Civil Engineering, Indian Institute of Science, Bangalore 560012 (India); Vasu, Ram Mohan [Department of Instrumentation and Applied Physics, Indian Institute of Science, Bangalore 560012 (India)
2014-08-01
A new global stochastic search, guided mainly through derivative-free directional information computable from the sample statistical moments of the design variables within a Monte Carlo setup, is proposed. The search is aided by imparting to the directional update term additional layers of random perturbations referred to as ‘coalescence’ and ‘scrambling’. A selection step, constituting yet another avenue for random perturbation, completes the global search. The direction-driven nature of the search is manifest in the local extremization and coalescence components, which are posed as martingale problems that yield gain-like update terms upon discretization. As anticipated and numerically demonstrated, to a limited extent, against the problem of parameter recovery given the chaotic response histories of a couple of nonlinear oscillators, the proposed method appears to offer a more rational, more accurate and faster alternative to most available evolutionary schemes, prominently the particle swarm optimization. - Highlights: • Evolutionary global optimization is posed as a perturbed martingale problem. • Resulting search via additive updates is a generalization over Gateaux derivatives. • Additional layers of random perturbation help avoid trapping at local extrema. • The approach ensures efficient design space exploration and high accuracy. • The method is numerically assessed via parameter recovery of chaotic oscillators.
SU-E-I-23: A General KV Constrained Optimization of CNR for CT Abdominal Imaging
International Nuclear Information System (INIS)
Weir, V; Zhang, J
2015-01-01
Purpose: While Tube current modulation has been well accepted for CT dose reduction, kV adjusting in clinical settings is still at its early stage. This is mainly due to the limited kV options of most current CT scanners. kV adjusting can potentially reduce radiation dose and optimize image quality. This study is to optimize CT abdomen imaging acquisition based on the assumption of a continuous kV, with the goal to provide the best contrast to noise ratio (CNR). Methods: For a given dose (CTDIvol) level, the CNRs at different kV and pitches were measured with an ACR GAMMEX phantom. The phantom was scanned in a Siemens Sensation 64 scanner and a GE VCT 64 scanner. A constrained mathematical optimization was used to find the kV which led to the highest CNR for the anatomy and pitch setting. Parametric equations were obtained from polynomial fitting of plots of kVs vs CNRs. A suitable constraint region for optimization was chosen. Subsequent optimization yielded a peak CNR at a particular kV for different collimations and pitch setting. Results: The constrained mathematical optimization approach yields kV of 114.83 and 113.46, with CNRs of 1.27 and 1.11 at the pitch of 1.2 and 1.4, respectively, for the Siemens Sensation 64 scanner with the collimation of 32 x 0.625mm. An optimized kV of 134.25 and 1.51 CNR is obtained for a GE VCT 64 slice scanner with a collimation of 32 x 0.625mm and a pitch of 0.969. At 0.516 pitch and 32 x 0.625 mm an optimized kV of 133.75 and a CNR of 1.14 was found for the GE VCT 64 slice scanner. Conclusion: CNR in CT image acquisition can be further optimized with a continuous kV option instead of current discrete or fixed kV settings. A continuous kV option is a key for individualized CT protocols
SU-E-I-23: A General KV Constrained Optimization of CNR for CT Abdominal Imaging
Energy Technology Data Exchange (ETDEWEB)
Weir, V; Zhang, J [University of Kentucky, Lexington, KY (United States)
2015-06-15
Purpose: While Tube current modulation has been well accepted for CT dose reduction, kV adjusting in clinical settings is still at its early stage. This is mainly due to the limited kV options of most current CT scanners. kV adjusting can potentially reduce radiation dose and optimize image quality. This study is to optimize CT abdomen imaging acquisition based on the assumption of a continuous kV, with the goal to provide the best contrast to noise ratio (CNR). Methods: For a given dose (CTDIvol) level, the CNRs at different kV and pitches were measured with an ACR GAMMEX phantom. The phantom was scanned in a Siemens Sensation 64 scanner and a GE VCT 64 scanner. A constrained mathematical optimization was used to find the kV which led to the highest CNR for the anatomy and pitch setting. Parametric equations were obtained from polynomial fitting of plots of kVs vs CNRs. A suitable constraint region for optimization was chosen. Subsequent optimization yielded a peak CNR at a particular kV for different collimations and pitch setting. Results: The constrained mathematical optimization approach yields kV of 114.83 and 113.46, with CNRs of 1.27 and 1.11 at the pitch of 1.2 and 1.4, respectively, for the Siemens Sensation 64 scanner with the collimation of 32 x 0.625mm. An optimized kV of 134.25 and 1.51 CNR is obtained for a GE VCT 64 slice scanner with a collimation of 32 x 0.625mm and a pitch of 0.969. At 0.516 pitch and 32 x 0.625 mm an optimized kV of 133.75 and a CNR of 1.14 was found for the GE VCT 64 slice scanner. Conclusion: CNR in CT image acquisition can be further optimized with a continuous kV option instead of current discrete or fixed kV settings. A continuous kV option is a key for individualized CT protocols.
Dual Schroedinger Equation as Global Optimization Algorithm
International Nuclear Information System (INIS)
Huang Xiaofei; eGain Communications, Mountain View, CA 94043
2011-01-01
The dual Schroedinger equation is defined as replacing the imaginary number i by -1 in the original one. This paper shows that the dual equation shares the same stationary states as the original one. Different from the original one, it explicitly defines a dynamic process for a system to evolve from any state to lower energy states and eventually to the lowest one. Its power as a global optimization algorithm might be used by nature for constructing atoms and molecules. It shall be interesting to verify its existence in nature.
Constraining the global carbon budget from global to regional scales - The measurement challenge
International Nuclear Information System (INIS)
Francey, R.J.; Rayner, P.J.; Allison, C.E.
2002-01-01
The Global Carbon Cycle can be modelled by a Bayesian synthesis inversion technique, where measured atmospheric CO 2 concentrations and isotopic compositions are analysed by use of an atmospheric transport model and estimates of regional sources and sinks of atmospheric carbon. The uncertainty associated to carbon flux estimates even on a regional scale can be improved considerably using the inversion technique. In this approach, besides the necessary control on the precision of atmospheric transport models and on the constraints for surface fluxes, an important component is the calibration of atmospheric CO 2 concentration and isotope measurements. The recent improved situation in respect to data comparability is discussed using results of conducted interlaboratory comparison exercises and larger scale calibration programs are proposed for the future to further improve the comparability of analytical data. (author)
A simple two stage optimization algorithm for constrained power economic dispatch
International Nuclear Information System (INIS)
Huang, G.; Song, K.
1994-01-01
A simple two stage optimization algorithm is proposed and investigated for fast computation of constrained power economic dispatch control problems. The method is a simple demonstration of the hierarchical aggregation-disaggregation (HAD) concept. The algorithm first solves an aggregated problem to obtain an initial solution. This aggregated problem turns out to be classical economic dispatch formulation, and it can be solved in 1% of overall computation time. In the second stage, linear programming method finds optimal solution which satisfies power balance constraints, generation and transmission inequality constraints and security constraints. Implementation of the algorithm for IEEE systems and EPRI Scenario systems shows that the two stage method obtains average speedup ratio 10.64 as compared to classical LP-based method
Reinforcement learning solution for HJB equation arising in constrained optimal control problem.
Luo, Biao; Wu, Huai-Ning; Huang, Tingwen; Liu, Derong
2015-11-01
The constrained optimal control problem depends on the solution of the complicated Hamilton-Jacobi-Bellman equation (HJBE). In this paper, a data-based off-policy reinforcement learning (RL) method is proposed, which learns the solution of the HJBE and the optimal control policy from real system data. One important feature of the off-policy RL is that its policy evaluation can be realized with data generated by other behavior policies, not necessarily the target policy, which solves the insufficient exploration problem. The convergence of the off-policy RL is proved by demonstrating its equivalence to the successive approximation approach. Its implementation procedure is based on the actor-critic neural networks structure, where the function approximation is conducted with linearly independent basis functions. Subsequently, the convergence of the implementation procedure with function approximation is also proved. Finally, its effectiveness is verified through computer simulations. Copyright © 2015 Elsevier Ltd. All rights reserved.
Risk-Constrained Dynamic Programming for Optimal Mars Entry, Descent, and Landing
Ono, Masahiro; Kuwata, Yoshiaki
2013-01-01
A chance-constrained dynamic programming algorithm was developed that is capable of making optimal sequential decisions within a user-specified risk bound. This work handles stochastic uncertainties over multiple stages in the CEMAT (Combined EDL-Mobility Analyses Tool) framework. It was demonstrated by a simulation of Mars entry, descent, and landing (EDL) using real landscape data obtained from the Mars Reconnaissance Orbiter. Although standard dynamic programming (DP) provides a general framework for optimal sequential decisionmaking under uncertainty, it typically achieves risk aversion by imposing an arbitrary penalty on failure states. Such a penalty-based approach cannot explicitly bound the probability of mission failure. A key idea behind the new approach is called risk allocation, which decomposes a joint chance constraint into a set of individual chance constraints and distributes risk over them. The joint chance constraint was reformulated into a constraint on an expectation over a sum of an indicator function, which can be incorporated into the cost function by dualizing the optimization problem. As a result, the chance-constraint optimization problem can be turned into an unconstrained optimization over a Lagrangian, which can be solved efficiently using a standard DP approach.
Adaptively Constrained Stochastic Model Predictive Control for the Optimal Dispatch of Microgrid
Directory of Open Access Journals (Sweden)
Xiaogang Guo
2018-01-01
Full Text Available In this paper, an adaptively constrained stochastic model predictive control (MPC is proposed to achieve less-conservative coordination between energy storage units and uncertain renewable energy sources (RESs in a microgrid (MG. Besides the economic objective of MG operation, the limits of state-of-charge (SOC and discharging/charging power of the energy storage unit are formulated as chance constraints when accommodating uncertainties of RESs, considering mild violations of these constraints are allowed during long-term operation, and a closed-loop online update strategy is performed to adaptively tighten or relax constraints according to the actual deviation probability of violation level from the desired one as well as the current change rate of deviation probability. Numerical studies show that the proposed adaptively constrained stochastic MPC for MG optimal operation is much less conservative compared with the scenario optimization based robust MPC, and also presents a better convergence performance to the desired constraint violation level than other online update strategies.
A branch and bound algorithm for the global optimization of Hessian Lipschitz continuous functions
Fowkes, Jaroslav M.
2012-06-21
We present a branch and bound algorithm for the global optimization of a twice differentiable nonconvex objective function with a Lipschitz continuous Hessian over a compact, convex set. The algorithm is based on applying cubic regularisation techniques to the objective function within an overlapping branch and bound algorithm for convex constrained global optimization. Unlike other branch and bound algorithms, lower bounds are obtained via nonconvex underestimators of the function. For a numerical example, we apply the proposed branch and bound algorithm to radial basis function approximations. © 2012 Springer Science+Business Media, LLC.
Global Optimization using Interval Analysis : Interval Optimization for Aerospace Applications
Van Kampen, E.
2010-01-01
Optimization is an important element in aerospace related research. It is encountered for example in trajectory optimization problems, such as: satellite formation flying, spacecraft re-entry optimization and airport approach and departure optimization; in control optimization, for example in
Directory of Open Access Journals (Sweden)
Charles Tatkeu
2008-12-01
Full Text Available We propose a global convergence baud-spaced blind equalization method in this paper. This method is based on the application of both generalized pattern optimization and channel surfing reinitialization. The potentially used unimodal cost function relies on higher- order statistics, and its optimization is achieved using a pattern search algorithm. Since the convergence to the global minimum is not unconditionally warranted, we make use of channel surfing reinitialization (CSR strategy to find the right global minimum. The proposed algorithm is analyzed, and simulation results using a severe frequency selective propagation channel are given. Detailed comparisons with constant modulus algorithm (CMA are highlighted. The proposed algorithm performances are evaluated in terms of intersymbol interference, normalized received signal constellations, and root mean square error vector magnitude. In case of nonconstant modulus input signals, our algorithm outperforms significantly CMA algorithm with full channel surfing reinitialization strategy. However, comparable performances are obtained for constant modulus signals.
Zaouche, Abdelouahib; Dayoub, Iyad; Rouvaen, Jean Michel; Tatkeu, Charles
2008-12-01
We propose a global convergence baud-spaced blind equalization method in this paper. This method is based on the application of both generalized pattern optimization and channel surfing reinitialization. The potentially used unimodal cost function relies on higher- order statistics, and its optimization is achieved using a pattern search algorithm. Since the convergence to the global minimum is not unconditionally warranted, we make use of channel surfing reinitialization (CSR) strategy to find the right global minimum. The proposed algorithm is analyzed, and simulation results using a severe frequency selective propagation channel are given. Detailed comparisons with constant modulus algorithm (CMA) are highlighted. The proposed algorithm performances are evaluated in terms of intersymbol interference, normalized received signal constellations, and root mean square error vector magnitude. In case of nonconstant modulus input signals, our algorithm outperforms significantly CMA algorithm with full channel surfing reinitialization strategy. However, comparable performances are obtained for constant modulus signals.
From the ground up: global nitrous oxide sources are constrained by stable isotope values.
Directory of Open Access Journals (Sweden)
David M Snider
Full Text Available Rising concentrations of nitrous oxide (N2O in the atmosphere are causing widespread concern because this trace gas plays a key role in the destruction of stratospheric ozone and it is a strong greenhouse gas. The successful mitigation of N2O emissions requires a solid understanding of the relative importance of all N2O sources and sinks. Stable isotope ratio measurements (δ15N-N2O and δ18O-N2O, including the intramolecular distribution of 15N (site preference, are one way to track different sources if they are isotopically distinct. 'Top-down' isotope mass-balance studies have had limited success balancing the global N2O budget thus far because the isotopic signatures of soil, freshwater, and marine sources are poorly constrained and a comprehensive analysis of global N2O stable isotope measurements has not been done. Here we used a robust analysis of all available in situ measurements to define key global N2O sources. We showed that the marine source is isotopically distinct from soil and freshwater N2O (the continental source. Further, the global average source (sum of all natural and anthropogenic sources is largely controlled by soils and freshwaters. These findings substantiate past modelling studies that relied on several assumptions about the global N2O cycle. Finally, a two-box-model and a Bayesian isotope mixing model revealed marine and continental N2O sources have relative contributions of 24-26% and 74-76% to the total, respectively. Further, the Bayesian modeling exercise indicated the N2O flux from freshwaters may be much larger than currently thought.
Preventive Security-Constrained Optimal Power Flow Considering UPFC Control Modes
Directory of Open Access Journals (Sweden)
Xi Wu
2017-08-01
Full Text Available The successful application of the unified power flow controller (UPFC provides a new control method for the secure and economic operation of power system. In order to make the full use of UPFC and improve the economic efficiency and static security of a power system, a preventive security-constrained power flow optimization method considering UPFC control modes is proposed in this paper. Firstly, an iterative method considering UPFC control modes is deduced for power flow calculation. Taking into account the influence of different UPFC control modes on the distribution of power flow after N-1 contingency, the optimization model is then constructed by setting a minimal system operation cost and a maximum static security margin as the objective. Based on this model, the particle swarm optimization (PSO algorithm is utilized to optimize power system operating parameters and UPFC control modes simultaneously. Finally, a standard IEEE 30-bus system is utilized to demonstrate that the proposed method fully exploits the potential of static control of UPFC and significantly increases the economic efficiency and static security of the power system.
Luo, Biao; Liu, Derong; Wu, Huai-Ning
2018-06-01
Reinforcement learning has proved to be a powerful tool to solve optimal control problems over the past few years. However, the data-based constrained optimal control problem of nonaffine nonlinear discrete-time systems has rarely been studied yet. To solve this problem, an adaptive optimal control approach is developed by using the value iteration-based Q-learning (VIQL) with the critic-only structure. Most of the existing constrained control methods require the use of a certain performance index and only suit for linear or affine nonlinear systems, which is unreasonable in practice. To overcome this problem, the system transformation is first introduced with the general performance index. Then, the constrained optimal control problem is converted to an unconstrained optimal control problem. By introducing the action-state value function, i.e., Q-function, the VIQL algorithm is proposed to learn the optimal Q-function of the data-based unconstrained optimal control problem. The convergence results of the VIQL algorithm are established with an easy-to-realize initial condition . To implement the VIQL algorithm, the critic-only structure is developed, where only one neural network is required to approximate the Q-function. The converged Q-function obtained from the critic-only VIQL method is employed to design the adaptive constrained optimal controller based on the gradient descent scheme. Finally, the effectiveness of the developed adaptive control method is tested on three examples with computer simulation.
A one-layer recurrent neural network for constrained nonsmooth optimization.
Liu, Qingshan; Wang, Jun
2011-10-01
This paper presents a novel one-layer recurrent neural network modeled by means of a differential inclusion for solving nonsmooth optimization problems, in which the number of neurons in the proposed neural network is the same as the number of decision variables of optimization problems. Compared with existing neural networks for nonsmooth optimization problems, the global convexity condition on the objective functions and constraints is relaxed, which allows the objective functions and constraints to be nonconvex. It is proven that the state variables of the proposed neural network are convergent to optimal solutions if a single design parameter in the model is larger than a derived lower bound. Numerical examples with simulation results substantiate the effectiveness and illustrate the characteristics of the proposed neural network.
Massioni, Paolo; Massari, Mauro
2018-05-01
This paper describes an interesting and powerful approach to the constrained fuel-optimal control of spacecraft in close relative motion. The proposed approach is well suited for problems under linear dynamic equations, therefore perfectly fitting to the case of spacecraft flying in close relative motion. If the solution of the optimisation is approximated as a polynomial with respect to the time variable, then the problem can be approached with a technique developed in the control engineering community, known as "Sum Of Squares" (SOS), and the constraints can be reduced to bounds on the polynomials. Such a technique allows rewriting polynomial bounding problems in the form of convex optimisation problems, at the cost of a certain amount of conservatism. The principles of the techniques are explained and some application related to spacecraft flying in close relative motion are shown.
A Constrained Least Squares Approach to Mobile Positioning: Algorithms and Optimality
Cheung, KW; So, HC; Ma, W.-K.; Chan, YT
2006-12-01
The problem of locating a mobile terminal has received significant attention in the field of wireless communications. Time-of-arrival (TOA), received signal strength (RSS), time-difference-of-arrival (TDOA), and angle-of-arrival (AOA) are commonly used measurements for estimating the position of the mobile station. In this paper, we present a constrained weighted least squares (CWLS) mobile positioning approach that encompasses all the above described measurement cases. The advantages of CWLS include performance optimality and capability of extension to hybrid measurement cases (e.g., mobile positioning using TDOA and AOA measurements jointly). Assuming zero-mean uncorrelated measurement errors, we show by mean and variance analysis that all the developed CWLS location estimators achieve zero bias and the Cramér-Rao lower bound approximately when measurement error variances are small. The asymptotic optimum performance is also confirmed by simulation results.
Directory of Open Access Journals (Sweden)
Kai Yang
2016-01-01
Full Text Available This work investigates a bioinspired microimmune optimization algorithm to solve a general kind of single-objective nonlinear constrained expected value programming without any prior distribution. In the study of algorithm, two lower bound sample estimates of random variables are theoretically developed to estimate the empirical values of individuals. Two adaptive racing sampling schemes are designed to identify those competitive individuals in a given population, by which high-quality individuals can obtain large sampling size. An immune evolutionary mechanism, along with a local search approach, is constructed to evolve the current population. The comparative experiments have showed that the proposed algorithm can effectively solve higher-dimensional benchmark problems and is of potential for further applications.
Optimal Financing Decisions of Two Cash-Constrained Supply Chains with Complementary Products
Directory of Open Access Journals (Sweden)
Yuting Li
2016-04-01
Full Text Available In recent years; financing difficulties have been obsessed small and medium enterprises (SMEs; especially emerging SMEs. Inter-members’ joint financing within a supply chain is one of solutions for SMEs. How about members’ joint financing of inter-supply chains? In order to answer the question, we firstly employ the Stackelberg game to propose three kinds of financing decision models of two cash-constrained supply chains with complementary products. Secondly, we analyze qualitatively these models and find the joint financing decision of the two supply chains is the most optimal one. Lastly, we conduct some numerical simulations not only to illustrate above results but also to find that the larger are cross-price sensitivity coefficients; the higher is the motivation for participants to make joint financing decisions; and the more are profits for them to gain.
Sarghini, Fabrizio; De Vivo, Angela; Marra, Francesco
2017-10-01
Computational science and engineering methods have allowed a major change in the way products and processes are designed, as validated virtual models - capable to simulate physical, chemical and bio changes occurring during production processes - can be realized and used in place of real prototypes and performing experiments, often time and money consuming. Among such techniques, Optimal Shape Design (OSD) (Mohammadi & Pironneau, 2004) represents an interesting approach. While most classical numerical simulations consider fixed geometrical configurations, in OSD a certain number of geometrical degrees of freedom is considered as a part of the unknowns: this implies that the geometry is not completely defined, but part of it is allowed to move dynamically in order to minimize or maximize the objective function. The applications of optimal shape design (OSD) are uncountable. For systems governed by partial differential equations, they range from structure mechanics to electromagnetism and fluid mechanics or to a combination of the three. This paper presents one of possible applications of OSD, particularly how extrusion bell shape, for past production, can be designed by applying a multivariate constrained shape optimization.
Ma, Jun; Chen, Si-Lu; Kamaldin, Nazir; Teo, Chek Sing; Tay, Arthur; Mamun, Abdullah Al; Tan, Kok Kiong
2017-11-01
The biaxial gantry is widely used in many industrial processes that require high precision Cartesian motion. The conventional rigid-link version suffers from breaking down of joints if any de-synchronization between the two carriages occurs. To prevent above potential risk, a flexure-linked biaxial gantry is designed to allow a small rotation angle of the cross-arm. Nevertheless, the chattering of control signals and inappropriate design of the flexure joint will possibly induce resonant modes of the end-effector. Thus, in this work, the design requirements in terms of tracking accuracy, biaxial synchronization, and resonant mode suppression are achieved by integrated optimization of the stiffness of flexures and PID controller parameters for a class of point-to-point reference trajectories with same dynamics but different steps. From here, an H 2 optimization problem with defined constraints is formulated, and an efficient iterative solver is proposed by hybridizing direct computation of constrained projection gradient and line search of optimal step. Comparative experimental results obtained on the testbed are presented to verify the effectiveness of the proposed method. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
International Nuclear Information System (INIS)
Park, T. K.; Joo, H. G.; Kim, C. H.
2010-01-01
In order to find the most economical loading pattern (LP) considering multi-cycle fuel loading, multi-objective fuel LP optimization problems are examined by employing an adaptively constrained discontinuous penalty function (ACDPF) method. This is an improved method to simplify the complicated acceptance logic of the original DPF method in that the stochastic effects caused by the different random number sequence can be reduced. The effectiveness of the multi-objective simulated annealing (SA) algorithm employing ACDPF is examined for the reload core LP of Cycle 4 of Yonggwang Nuclear Unit 4. Several optimization runs are performed with different numbers of objectives consisting of cycle length and average burnup of fuels to be discharged or reloaded. The candidate LPs obtained from the multi-objective optimization runs turn out to be better than the reference LP in the aspects of cycle length and utilization of given fuels. It is note that the proposed ACDPF based MOSA algorithm can be a practical method to obtain an economical LP considering multi-cycle fuel loading. (authors)
Stochastic Averaging for Constrained Optimization With Application to Online Resource Allocation
Chen, Tianyi; Mokhtari, Aryan; Wang, Xin; Ribeiro, Alejandro; Giannakis, Georgios B.
2017-06-01
Existing approaches to resource allocation for nowadays stochastic networks are challenged to meet fast convergence and tolerable delay requirements. The present paper leverages online learning advances to facilitate stochastic resource allocation tasks. By recognizing the central role of Lagrange multipliers, the underlying constrained optimization problem is formulated as a machine learning task involving both training and operational modes, with the goal of learning the sought multipliers in a fast and efficient manner. To this end, an order-optimal offline learning approach is developed first for batch training, and it is then generalized to the online setting with a procedure termed learn-and-adapt. The novel resource allocation protocol permeates benefits of stochastic approximation and statistical learning to obtain low-complexity online updates with learning errors close to the statistical accuracy limits, while still preserving adaptation performance, which in the stochastic network optimization context guarantees queue stability. Analysis and simulated tests demonstrate that the proposed data-driven approach improves the delay and convergence performance of existing resource allocation schemes.
Modares, Hamidreza; Lewis, Frank L; Naghibi-Sistani, Mohammad-Bagher
2013-10-01
This paper presents an online policy iteration (PI) algorithm to learn the continuous-time optimal control solution for unknown constrained-input systems. The proposed PI algorithm is implemented on an actor-critic structure where two neural networks (NNs) are tuned online and simultaneously to generate the optimal bounded control policy. The requirement of complete knowledge of the system dynamics is obviated by employing a novel NN identifier in conjunction with the actor and critic NNs. It is shown how the identifier weights estimation error affects the convergence of the critic NN. A novel learning rule is developed to guarantee that the identifier weights converge to small neighborhoods of their ideal values exponentially fast. To provide an easy-to-check persistence of excitation condition, the experience replay technique is used. That is, recorded past experiences are used simultaneously with current data for the adaptation of the identifier weights. Stability of the whole system consisting of the actor, critic, system state, and system identifier is guaranteed while all three networks undergo adaptation. Convergence to a near-optimal control law is also shown. The effectiveness of the proposed method is illustrated with a simulation example.
International Nuclear Information System (INIS)
Hinze, J F; Klein, S A; Nellis, G F
2015-01-01
Mixed refrigerant (MR) working fluids can significantly increase the cooling capacity of a Joule-Thomson (JT) cycle. The optimization of MRJT systems has been the subject of substantial research. However, most optimization techniques do not model the recuperator in sufficient detail. For example, the recuperator is usually assumed to have a heat transfer coefficient that does not vary with the mixture. Ongoing work at the University of Wisconsin-Madison has shown that the heat transfer coefficients for two-phase flow are approximately three times greater than for a single phase mixture when the mixture quality is between 15% and 85%. As a result, a system that optimizes a MR without also requiring that the flow be in this quality range may require an extremely large recuperator or not achieve the performance predicted by the model. To ensure optimal performance of the JT cycle, the MR should be selected such that it is entirely two-phase within the recuperator. To determine the optimal MR composition, a parametric study was conducted assuming a thermodynamically ideal cycle. The results of the parametric study are graphically presented on a contour plot in the parameter space consisting of the extremes of the qualities that exist within the recuperator. The contours show constant values of the normalized refrigeration power. This ‘map’ shows the effect of MR composition on the cycle performance and it can be used to select the MR that provides a high cooling load while also constraining the recuperator to be two phase. The predicted best MR composition can be used as a starting point for experimentally determining the best MR. (paper)
DATA-CONSTRAINED CORONAL MASS EJECTIONS IN A GLOBAL MAGNETOHYDRODYNAMICS MODEL
Energy Technology Data Exchange (ETDEWEB)
Jin, M. [Lockheed Martin Solar and Astrophysics Lab, Palo Alto, CA 94304 (United States); Manchester, W. B.; Van der Holst, B.; Sokolov, I.; Tóth, G.; Gombosi, T. I. [Climate and Space Sciences and Engineering, University of Michigan, Ann Arbor, MI 48109 (United States); Mullinix, R. E.; Taktakishvili, A.; Chulaki, A., E-mail: jinmeng@lmsal.com, E-mail: chipm@umich.edu, E-mail: richard.e.mullinix@nasa.gov, E-mail: Aleksandre.Taktakishvili-1@nasa.gov [Community Coordinated Modeling Center, NASA Goddard Space Flight Center, Greenbelt, MD 20771 (United States)
2017-01-10
We present a first-principles-based coronal mass ejection (CME) model suitable for both scientific and operational purposes by combining a global magnetohydrodynamics (MHD) solar wind model with a flux-rope-driven CME model. Realistic CME events are simulated self-consistently with high fidelity and forecasting capability by constraining initial flux rope parameters with observational data from GONG, SOHO /LASCO, and STEREO /COR. We automate this process so that minimum manual intervention is required in specifying the CME initial state. With the newly developed data-driven Eruptive Event Generator using Gibson–Low configuration, we present a method to derive Gibson–Low flux rope parameters through a handful of observational quantities so that the modeled CMEs can propagate with the desired CME speeds near the Sun. A test result with CMEs launched with different Carrington rotation magnetograms is shown. Our study shows a promising result for using the first-principles-based MHD global model as a forecasting tool, which is capable of predicting the CME direction of propagation, arrival time, and ICME magnetic field at 1 au (see the companion paper by Jin et al. 2016a).
Directory of Open Access Journals (Sweden)
Qingyang Zhang
2015-02-01
Full Text Available Bird Mating Optimizer (BMO is a novel meta-heuristic optimization algorithm inspired by intelligent mating behavior of birds. However, it is still insufficient in convergence of speed and quality of solution. To overcome these drawbacks, this paper proposes a hybrid algorithm (TLBMO, which is established by combining the advantages of Teaching-learning-based optimization (TLBO and Bird Mating Optimizer (BMO. The performance of TLBMO is evaluated on 23 benchmark functions, and compared with seven state-of-the-art approaches, namely BMO, TLBO, Artificial Bee Bolony (ABC, Particle Swarm Optimization (PSO, Fast Evolution Programming (FEP, Differential Evolution (DE, Group Search Optimization (GSO. Experimental results indicate that the proposed method performs better than other existing algorithms for global numerical optimization.
Directory of Open Access Journals (Sweden)
Narinder Singh
2018-03-01
Full Text Available The quest for an efficient nature-inspired optimization technique has continued over the last few decades. In this paper, a hybrid nature-inspired optimization technique has been proposed. The hybrid algorithm has been constructed using Mean Grey Wolf Optimizer (MGWO and Whale Optimizer Algorithm (WOA. We have utilized the spiral equation of Whale Optimizer Algorithm for two procedures in the Hybrid Approach GWO (HAGWO algorithm: (i firstly, we used the spiral equation in Grey Wolf Optimizer algorithm for balance between the exploitation and the exploration process in the new hybrid approach; and (ii secondly, we also applied this equation in the whole population in order to refrain from the premature convergence and trapping in local minima. The feasibility and effectiveness of the hybrid algorithm have been tested by solving some standard benchmarks, XOR, Baloon, Iris, Breast Cancer, Welded Beam Design, Pressure Vessel Design problems and comparing the results with those obtained through other metaheuristics. The solutions prove that the newly existing hybrid variant has higher stronger stability, faster convergence rate and computational accuracy than other nature-inspired metaheuristics on the maximum number of problems and can successfully resolve the function of constrained nonlinear optimization in reality.
Asgari, Jamal; Mohammadloo, Tannaz H.; Amiri-Simkooei, Ali Reza
2015-09-01
GNSS kinematic techniques are capable of providing precise coordinates in extremely short observation time-span. These methods usually determine the coordinates of an unknown station with respect to a reference one. To enhance the precision, accuracy, reliability and integrity of the estimated unknown parameters, GNSS kinematic equations are to be augmented by possible constraints. Such constraints could be derived from the geometric relation of the receiver positions in motion. This contribution presents the formulation of the constrained kinematic global navigation satellite systems positioning. Constraints effectively restrict the definition domain of the unknown parameters from the three-dimensional space to a subspace defined by the equation of motion. To test the concept of the constrained kinematic positioning method, the equation of a circle is employed as a constraint. A device capable of moving on a circle was made and the observations from 11 positions on the circle were analyzed. Relative positioning was conducted by considering the center of the circle as the reference station. The equation of the receiver's motion was rewritten in the ECEF coordinates system. A special attention is drawn onto how a constraint is applied to kinematic positioning. Implementing the constraint in the positioning process provides much more precise results compared to the unconstrained case. This has been verified based on the results obtained from the covariance matrix of the estimated parameters and the empirical results using kinematic positioning samples as well. The theoretical standard deviations of the horizontal components are reduced by a factor ranging from 1.24 to 2.64. The improvement on the empirical standard deviation of the horizontal components ranges from 1.08 to 2.2.
Stress-constrained truss topology optimization problems that can be solved by linear programming
DEFF Research Database (Denmark)
Stolpe, Mathias; Svanberg, Krister
2004-01-01
We consider the problem of simultaneously selecting the material and determining the area of each bar in a truss structure in such a way that the cost of the structure is minimized subject to stress constraints under a single load condition. We show that such problems can be solved by linear...... programming to give the global optimum, and that two different materials are always sufficient in an optimal structure....
Parallel Global Optimization with the Particle Swarm Algorithm (Preprint)
National Research Council Canada - National Science Library
Schutte, J. F; Reinbolt, J. A; Fregly, B. J; Haftka, R. T; George, A. D
2004-01-01
.... To obtain enhanced computational throughput and global search capability, we detail the coarse-grained parallelization of an increasingly popular global search method, the Particle Swarm Optimization (PSO) algorithm...
Mini-batch optimized full waveform inversion with geological constrained gradient filtering
Yang, Hui; Jia, Junxiong; Wu, Bangyu; Gao, Jinghuai
2018-05-01
High computation cost and generating solutions without geological sense have hindered the wide application of Full Waveform Inversion (FWI). Source encoding technique is a way to dramatically reduce the cost of FWI but subject to fix-spread acquisition setup requirement and slow convergence for the suppression of cross-talk. Traditionally, gradient regularization or preconditioning is applied to mitigate the ill-posedness. An isotropic smoothing filter applied on gradients generally gives non-geological inversion results, and could also introduce artifacts. In this work, we propose to address both the efficiency and ill-posedness of FWI by a geological constrained mini-batch gradient optimization method. The mini-batch gradient descent optimization is adopted to reduce the computation time by choosing a subset of entire shots for each iteration. By jointly applying the structure-oriented smoothing to the mini-batch gradient, the inversion converges faster and gives results with more geological meaning. Stylized Marmousi model is used to show the performance of the proposed method on realistic synthetic model.
Liu, Qiang; Chattopadhyay, Aditi
2000-06-01
Aeromechanical stability plays a critical role in helicopter design and lead-lag damping is crucial to this design. In this paper, the use of segmented constrained damping layer (SCL) treatment and composite tailoring is investigated for improved rotor aeromechanical stability using formal optimization technique. The principal load-carrying member in the rotor blade is represented by a composite box beam, of arbitrary thickness, with surface bonded SCLs. A comprehensive theory is used to model the smart box beam. A ground resonance analysis model and an air resonance analysis model are implemented in the rotor blade built around the composite box beam with SCLs. The Pitt-Peters dynamic inflow model is used in air resonance analysis under hover condition. A hybrid optimization technique is used to investigate the optimum design of the composite box beam with surface bonded SCLs for improved damping characteristics. Parameters such as stacking sequence of the composite laminates and placement of SCLs are used as design variables. Detailed numerical studies are presented for aeromechanical stability analysis. It is shown that optimum blade design yields significant increase in rotor lead-lag regressive modal damping compared to the initial system.
Directory of Open Access Journals (Sweden)
Longfei He
2014-01-01
Full Text Available We focus on the joint production planning of complex supply chains facing stochastic demands and being constrained by carbon emission reduction policies. We pick two typical carbon emission reduction policies to research how emission regulation influences the profit and carbon footprint of a typical supply chain. We use the input-output model to capture the interrelated demand link between an arbitrary pair of two nodes in scenarios without or with carbon emission constraints. We design optimization algorithm to obtain joint optimal production quantities combination for maximizing overall profit under regulatory policies, respectively. Furthermore, numerical studies by featuring exponentially distributed demand compare systemwide performances in various scenarios. We build the “carbon emission elasticity of profit (CEEP” index as a metric to evaluate the impact of regulatory policies on both chainwide emissions and profit. Our results manifest that by facilitating the mandatory emission cap in proper installation within the network one can balance well effective emission reduction and associated acceptable profit loss. The outcome that CEEP index when implementing Carbon emission tax is elastic implies that the scale of profit loss is greater than that of emission reduction, which shows that this policy is less effective than mandatory cap from industry standpoint at least.
Ichii, K.; Kondo, M.; Wang, W.; Hashimoto, H.; Nemani, R. R.
2012-12-01
Various satellite-based spatial products such as evapotranspiration (ET) and gross primary productivity (GPP) are now produced by integration of ground and satellite observations. Effective use of these multiple satellite-based products in terrestrial biosphere models is an important step toward better understanding of terrestrial carbon and water cycles. However, due to the complexity of terrestrial biosphere models with large number of model parameters, the application of these spatial data sets in terrestrial biosphere models is difficult. In this study, we established an effective but simple framework to refine a terrestrial biosphere model, Biome-BGC, using multiple satellite-based products as constraints. We tested the framework in the monsoon Asia region covered by AsiaFlux observations. The framework is based on the hierarchical analysis (Wang et al. 2009) with model parameter optimization constrained by satellite-based spatial data. The Biome-BGC model is separated into several tiers to minimize the freedom of model parameter selections and maximize the independency from the whole model. For example, the snow sub-model is first optimized using MODIS snow cover product, followed by soil water sub-model optimized by satellite-based ET (estimated by an empirical upscaling method; Support Vector Regression (SVR) method; Yang et al. 2007), photosynthesis model optimized by satellite-based GPP (based on SVR method), and respiration and residual carbon cycle models optimized by biomass data. As a result of initial assessment, we found that most of default sub-models (e.g. snow, water cycle and carbon cycle) showed large deviations from remote sensing observations. However, these biases were removed by applying the proposed framework. For example, gross primary productivities were initially underestimated in boreal and temperate forest and overestimated in tropical forests. However, the parameter optimization scheme successfully reduced these biases. Our analysis
Optimal control landscape for the generation of unitary transformations with constrained dynamics
International Nuclear Information System (INIS)
Hsieh, Michael; Wu, Rebing; Rabitz, Herschel; Lidar, Daniel
2010-01-01
The reliable and precise generation of quantum unitary transformations is essential for the realization of a number of fundamental objectives, such as quantum control and quantum information processing. Prior work has explored the optimal control problem of generating such unitary transformations as a surface-optimization problem over the quantum control landscape, defined as a metric for realizing a desired unitary transformation as a function of the control variables. It was found that under the assumption of nondissipative and controllable dynamics, the landscape topology is trap free, which implies that any reasonable optimization heuristic should be able to identify globally optimal solutions. The present work is a control landscape analysis, which incorporates specific constraints in the Hamiltonian that correspond to certain dynamical symmetries in the underlying physical system. It is found that the presence of such symmetries does not destroy the trap-free topology. These findings expand the class of quantum dynamical systems on which control problems are intrinsically amenable to a solution by optimal control.
The global economic long-term potential of modern biomass in a climate-constrained world
Klein, David; Humpenöder, Florian; Bauer, Nico; Dietrich, Jan Philipp; Popp, Alexander; Bodirsky, Benjamin Leon; Bonsch, Markus; Lotze-Campen, Hermann
2014-07-01
Low-stabilization scenarios consistent with the 2 °C target project large-scale deployment of purpose-grown lignocellulosic biomass. In case a GHG price regime integrates emissions from energy conversion and from land-use/land-use change, the strong demand for bioenergy and the pricing of terrestrial emissions are likely to coincide. We explore the global potential of purpose-grown lignocellulosic biomass and ask the question how the supply prices of biomass depend on prices for greenhouse gas (GHG) emissions from the land-use sector. Using the spatially explicit global land-use optimization model MAgPIE, we construct bioenergy supply curves for ten world regions and a global aggregate in two scenarios, with and without a GHG tax. We find that the implementation of GHG taxes is crucial for the slope of the supply function and the GHG emissions from the land-use sector. Global supply prices start at 5 GJ-1 and increase almost linearly, doubling at 150 EJ (in 2055 and 2095). The GHG tax increases bioenergy prices by 5 GJ-1 in 2055 and by 10 GJ-1 in 2095, since it effectively stops deforestation and thus excludes large amounts of high-productivity land. Prices additionally increase due to costs for N2O emissions from fertilizer use. The GHG tax decreases global land-use change emissions by one-third. However, the carbon emissions due to bioenergy production increase by more than 50% from conversion of land that is not under emission control. Average yields required to produce 240 EJ in 2095 are roughly 600 GJ ha-1 yr-1 with and without tax.
The global economic long-term potential of modern biomass in a climate-constrained world
International Nuclear Information System (INIS)
Klein, David; Humpenöder, Florian; Bauer, Nico; Dietrich, Jan Philipp; Popp, Alexander; Leon Bodirsky, Benjamin; Bonsch, Markus; Lotze-Campen, Hermann
2014-01-01
Low-stabilization scenarios consistent with the 2 °C target project large-scale deployment of purpose-grown lignocellulosic biomass. In case a GHG price regime integrates emissions from energy conversion and from land-use/land-use change, the strong demand for bioenergy and the pricing of terrestrial emissions are likely to coincide. We explore the global potential of purpose-grown lignocellulosic biomass and ask the question how the supply prices of biomass depend on prices for greenhouse gas (GHG) emissions from the land-use sector. Using the spatially explicit global land-use optimization model MAgPIE, we construct bioenergy supply curves for ten world regions and a global aggregate in two scenarios, with and without a GHG tax. We find that the implementation of GHG taxes is crucial for the slope of the supply function and the GHG emissions from the land-use sector. Global supply prices start at $5 GJ −1 and increase almost linearly, doubling at 150 EJ (in 2055 and 2095). The GHG tax increases bioenergy prices by $5 GJ −1 in 2055 and by $10 GJ −1 in 2095, since it effectively stops deforestation and thus excludes large amounts of high-productivity land. Prices additionally increase due to costs for N 2 O emissions from fertilizer use. The GHG tax decreases global land-use change emissions by one-third. However, the carbon emissions due to bioenergy production increase by more than 50% from conversion of land that is not under emission control. Average yields required to produce 240 EJ in 2095 are roughly 600 GJ ha −1 yr −1 with and without tax. (letter)
Morgenthaler, George; Khatib, Nader; Kim, Byoungsoo
with information to improve their crop's vigor has been a major topic of interest. With world population growing exponentially, arable land being consumed by urbanization, and an unfavorable farm economy, the efficiency of farming must increase to meet future food requirements and to make farming a sustainable occupation for the farmer. "Precision Agriculture" refers to a farming methodology that applies nutrients and moisture only where and when they are needed in the field. The goal is to increase farm revenue by increasing crop yield and decreasing applications of costly chemical and water treatments. In addition, this methodology will decrease the environmental costs of farming, i.e., reduce air, soil, and water pollution. Sensing/Precision Agriculture has not grown as rapidly as early advocates envisioned. Technology for a successful Remote Sensing/Precision Agriculture system is now available. Commercial satellite systems can image (multi-spectral) the Earth with a resolution of approximately 2.5 m. Variable precision dispensing systems using GPS are available and affordable. Crop models that predict yield as a function of soil, chemical, and irrigation parameter levels have been formulated. Personal computers and internet access are in place in most farm homes and can provide a mechanism to periodically disseminate, e.g. bi-weekly, advice on what quantities of water and chemicals are needed in individual regions of the field. What is missing is a model that fuses the disparate sources of information on the current states of the crop and soil, and the remaining resource levels available with the decisions farmers are required to make. This must be a product that is easy for the farmer to understand and to implement. A "Constrained Optimization Feed-back Control Model" to fill this void will be presented. The objective function of the model will be used to maximize the farmer's profit by increasing yields while decreasing environmental costs and decreasing
Moreenthaler, George W.; Khatib, Nader; Kim, Byoungsoo
2003-08-01
For two decades now, the use of Remote Sensing/Precision Agriculture to improve farm yields while reducing the use of polluting chemicals and the limited water supply has been a major goal. With world population growing exponentially, arable land being consumed by urbanization, and an unfavorable farm economy, farm efficiency must increase to meet future food requirements and to make farming a sustainable, profitable occupation. "Precision Agriculture" refers to a farming methodology that applies nutrients and moisture only where and when they are needed in the field. The real goal is to increase farm profitability by identifying the additional treatments of chemicals and water that increase revenues more than they increase costs and do no exceed pollution standards (constrained optimization). Even though the economic and environmental benefits appear to be great, Remote Sensing/Precision Agriculture has not grown as rapidly as early advocates envisioned. Technology for a successful Remote Sensing/Precision Agriculture system is now in place, but other needed factors have been missing. Commercial satellite systems can now image the Earth (multi-spectrally) with a resolution as fine as 2.5 m. Precision variable dispensing systems using GPS are now available and affordable. Crop models that predict yield as a function of soil, chemical, and irrigation parameter levels have been developed. Personal computers and internet access are now in place in most farm homes and can provide a mechanism for periodically disseminating advice on what quantities of water and chemicals are needed in specific regions of each field. Several processes have been selected that fuse the disparate sources of information on the current and historic states of the crop and soil, and the remaining resource levels available, with the critical decisions that farmers are required to make. These are done in a way that is easy for the farmer to understand and profitable to implement. A "Constrained
Directory of Open Access Journals (Sweden)
Qiuyu Wang
2014-01-01
descent method at first finite number of steps and then by conjugate gradient method subsequently. Under some appropriate conditions, we show that the algorithm converges globally. Numerical experiments and comparisons by using some box-constrained problems from CUTEr library are reported. Numerical comparisons illustrate that the proposed method is promising and competitive with the well-known method—L-BFGS-B.
Constraining Cosmic Dawn and Cosmological Reionization via the global redshifted 21-cm signal
Singh, Saurabh
2018-01-01
The formation of first stars and consequent thermal evolution in baryons during Cosmic Dawn and the Epoch of Reionization (EoR) is poorly constrained. The 21-cm line transition of neutral hydrogen is one of the richest probes of the astrophysics during this era. The signal has the potential to reveal the nature and timing of the emergence of first stars, first light, and the consequent evolution in thermal and ionization state of the baryons.The detection of the global redshifted 21-cm signal, which represents the mean thermal history of the gas, is challenging since it is extremely faint and seen through orders of magnitude stronger contributions from Galactic and extragalactic foregrounds. Man-made terrestrial Radio Frequency Interference (RFI) and the exacting tolerances required on instrument systematics make the detection even more daunting.The design considerations for a precision spectral radiometer are first listed, and a comparison is made of different radiometer configurations, including short and zero baseline interferometers along with methods to enhance the response. We discuss the relative merits of different methods.We then describe SARAS 2, a spectral radiometer custom-designed for precision measurement of the global 21-cm signal. SARAS 2 has been designed to have a system transfer function and internal systematics – both multiplicative and additive – to be spectrally smooth so as to allow a separation of foregrounds and systematics from plausible and predicted global cosmological 21-cm signals. The algorithms for calibration and RFI mitigation are carefully developed so that they do not introduce spectral features that may confuse the detection of the 21-cm signal.We present the outcomes for cosmology from analysis of 60 hr observing with the radiometer deployed at the Timbaktu Collective in Southern India. The detailed analysis of the data reveals an RMS noise level of 11 mK, without being limited by systematic structures. The likelihood
On the optimal identification of tag sets in time-constrained RFID configurations.
Vales-Alonso, Javier; Bueno-Delgado, María Victoria; Egea-López, Esteban; Alcaraz, Juan José; Pérez-Mañogil, Juan Manuel
2011-01-01
In Radio Frequency Identification facilities the identification delay of a set of tags is mainly caused by the random access nature of the reading protocol, yielding a random identification time of the set of tags. In this paper, the cumulative distribution function of the identification time is evaluated using a discrete time Markov chain for single-set time-constrained passive RFID systems, namely those ones where a single group of tags is assumed to be in the reading area and only for a bounded time (sojourn time) before leaving. In these scenarios some tags in a set may leave the reader coverage area unidentified. The probability of this event is obtained from the cumulative distribution function of the identification time as a function of the sojourn time. This result provides a suitable criterion to minimize the probability of losing tags. Besides, an identification strategy based on splitting the set of tags in smaller subsets is also considered. Results demonstrate that there are optimal splitting configurations that reduce the overall identification time while keeping the same probability of losing tags.
SmartFix: Indoor Locating Optimization Algorithm for Energy-Constrained Wearable Devices
Directory of Open Access Journals (Sweden)
Xiaoliang Wang
2017-01-01
Full Text Available Indoor localization technology based on Wi-Fi has long been a hot research topic in the past decade. Despite numerous solutions, new challenges have arisen along with the trend of smart home and wearable computing. For example, power efficiency needs to be significantly improved for resource-constrained wearable devices, such as smart watch and wristband. For a Wi-Fi-based locating system, most of the energy consumption can be attributed to real-time radio scan; however, simply reducing radio data collection will cause a serious loss of locating accuracy because of unstable Wi-Fi signals. In this paper, we present SmartFix, an optimization algorithm for indoor locating based on Wi-Fi RSS. SmartFix utilizes user motion features, extracts characteristic value from history trajectory, and corrects deviation caused by unstable Wi-Fi signals. We implemented a prototype of SmartFix both on Moto 360 2nd-generation Smartwatch and on HTC One Smartphone. We conducted experiments both in a large open area and in an office hall. Experiment results demonstrate that average locating error is less than 2 meters for more than 80% cases, and energy consumption is only 30% of Wi-Fi fingerprinting method under the same experiment circumstances.
3rd World Congress on Global Optimization in Engineering & Science
Ruan, Ning; Xing, Wenxun; WCGO-III; Advances in Global Optimization
2015-01-01
This proceedings volume addresses advances in global optimization—a multidisciplinary research field that deals with the analysis, characterization, and computation of global minima and/or maxima of nonlinear, non-convex, and nonsmooth functions in continuous or discrete forms. The volume contains selected papers from the third biannual World Congress on Global Optimization in Engineering & Science (WCGO), held in the Yellow Mountains, Anhui, China on July 8-12, 2013. The papers fall into eight topical sections: mathematical programming; combinatorial optimization; duality theory; topology optimization; variational inequalities and complementarity problems; numerical optimization; stochastic models and simulation; and complex simulation and supply chain analysis.
4th International Conference on Frontiers in Global Optimization
Pardalos, Panos
2004-01-01
Global Optimization has emerged as one of the most exciting new areas of mathematical programming. Global optimization has received a wide attraction from many fields in the past few years, due to the success of new algorithms for addressing previously intractable problems from diverse areas such as computational chemistry and biology, biomedicine, structural optimization, computer sciences, operations research, economics, and engineering design and control. This book contains refereed invited papers submitted at the 4th international confer ence on Frontiers in Global Optimization held at Santorini, Greece during June 8-12, 2003. Santorini is one of the few sites of Greece, with wild beauty created by the explosion of a volcano which is in the middle of the gulf of the island. The mystic landscape with its numerous mult-extrema, was an inspiring location particularly for researchers working on global optimization. The three previous conferences on "Recent Advances in Global Opti mization", "State-of-the-...
Microwave tomography global optimization, parallelization and performance evaluation
Noghanian, Sima; Desell, Travis; Ashtari, Ali
2014-01-01
This book provides a detailed overview on the use of global optimization and parallel computing in microwave tomography techniques. The book focuses on techniques that are based on global optimization and electromagnetic numerical methods. The authors provide parallelization techniques on homogeneous and heterogeneous computing architectures on high performance and general purpose futuristic computers. The book also discusses the multi-level optimization technique, hybrid genetic algorithm and its application in breast cancer imaging.
Constraining the Antarctic contribution to global sea-level change: ANDRILL and beyond
Naish, Timothy
2016-04-01
Observations, models and paleoclimate reconstructions suggest that Antarctica's marine-based ice sheets behave in an unstable manner with episodes of rapid retreat in response to warming climate. Understanding the processes involved in this "marine ice sheet instability" is key for improving estimates of Antarctic ice sheet contribution to future sea-level rise. Another motivating factor is that far-field sea-level reconstructions and ice sheet models imply global mean sea level (GMSL) was up to 20m and 10m higher, respectively, compared with present day, during the interglacials of the warm Pliocene (~4-3Ma) and Late Pleistocene (at ~400ka and 125ka). This was when atmospheric CO2 was between 280 and 400ppm and global average surface temperatures were 1 to 3°C warmer, suggesting polar ice sheets are highly sensitive to relatively modest increases in climate forcing. Such magnitudes of GMSL rise not only require near complete melt of the Greenland Ice Sheet and the West Antarctic Ice Sheet, but a substantial retreat of marine-based sectors of East Antarctic Ice Sheet. Recent geological drilling initiatives on the continental margin of Antarctica from both ship- (e.g. IODP; International Ocean Discovery Program) and ice-based (e.g. ANDRILL/Antarctic Geological Drilling) platforms have provided evidence supporting retreat of marine-based ice. However, without direct access through the ice sheet to archives preserved within sub-glacial sedimentary basins, the volume and extent of ice sheet retreat during past interglacials cannot be directly constrained. Sediment cores have been successfully recovered from beneath ice shelves by the ANDRILL Program and ice streams by the WISSARD (Whillans Ice Stream Sub-glacial Access Research Drilling) Project. Together with the potential of the new RAID (Rapid Access Ice Drill) initiative, these demonstrate the technological feasibility of accessing the subglacial bed and deeper sedimentary archives. In this talk I will outline the
Guo, Weian; Li, Wuzhao; Zhang, Qun; Wang, Lei; Wu, Qidi; Ren, Hongliang
2014-11-01
In evolutionary algorithms, elites are crucial to maintain good features in solutions. However, too many elites can make the evolutionary process stagnate and cannot enhance the performance. This article employs particle swarm optimization (PSO) and biogeography-based optimization (BBO) to propose a hybrid algorithm termed biogeography-based particle swarm optimization (BPSO) which could make a large number of elites effective in searching optima. In this algorithm, the whole population is split into several subgroups; BBO is employed to search within each subgroup and PSO for the global search. Since not all the population is used in PSO, this structure overcomes the premature convergence in the original PSO. Time complexity analysis shows that the novel algorithm does not increase the time consumption. Fourteen numerical benchmarks and four engineering problems with constraints are used to test the BPSO. To better deal with constraints, a fuzzy strategy for the number of elites is investigated. The simulation results validate the feasibility and effectiveness of the proposed algorithm.
Roselyn, J. Preetha; Devaraj, D.; Dash, Subhransu Sekhar
2013-11-01
Voltage stability is an important issue in the planning and operation of deregulated power systems. The voltage stability problems is a most challenging one for the system operators in deregulated power systems because of the intense use of transmission line capabilities and poor regulation in market environment. This article addresses the congestion management problem avoiding offline transmission capacity limits related to voltage stability by considering Voltage Security Constrained Optimal Power Flow (VSCOPF) problem in deregulated environment. This article presents the application of Multi Objective Differential Evolution (MODE) algorithm to solve the VSCOPF problem in new competitive power systems. The maximum of L-index of the load buses is taken as the indicator of voltage stability and is incorporated in the Optimal Power Flow (OPF) problem. The proposed method in hybrid power market which also gives solutions to voltage stability problems by considering the generation rescheduling cost and load shedding cost which relieves the congestion problem in deregulated environment. The buses for load shedding are selected based on the minimum eigen value of Jacobian with respect to the load shed. In the proposed approach, real power settings of generators in base case and contingency cases, generator bus voltage magnitudes, real and reactive power demands of selected load buses using sensitivity analysis are taken as the control variables and are represented as the combination of floating point numbers and integers. DE/randSF/1/bin strategy scheme of differential evolution with self-tuned parameter which employs binomial crossover and difference vector based mutation is used for the VSCOPF problem. A fuzzy based mechanism is employed to get the best compromise solution from the pareto front to aid the decision maker. The proposed VSCOPF planning model is implemented on IEEE 30-bus system, IEEE 57 bus practical system and IEEE 118 bus system. The pareto optimal
Decentralized Control Using Global Optimization (DCGO) (Preprint)
National Research Council Canada - National Science Library
Flint, Matthew; Khovanova, Tanya; Curry, Michael
2007-01-01
The coordination of a team of distributed air vehicles requires a complex optimization, balancing limited communication bandwidths, non-instantaneous planning times and network delays, while at the...
Interactive Cosegmentation Using Global and Local Energy Optimization
Xingping Dong,; Jianbing Shen,; Shao, Ling; Yang, Ming-Hsuan
2015-01-01
We propose a novel interactive cosegmentation method using global and local energy optimization. The global energy includes two terms: 1) the global scribbled energy and 2) the interimage energy. The first one utilizes the user scribbles to build the Gaussian mixture model and improve the cosegmentation performance. The second one is a global constraint, which attempts to match the histograms of common objects. To minimize the local energy, we apply the spline regression to learn the smoothne...
Bassen, David M; Vilkhovoy, Michael; Minot, Mason; Butcher, Jonathan T; Varner, Jeffrey D
2017-01-25
Ensemble modeling is a promising approach for obtaining robust predictions and coarse grained population behavior in deterministic mathematical models. Ensemble approaches address model uncertainty by using parameter or model families instead of single best-fit parameters or fixed model structures. Parameter ensembles can be selected based upon simulation error, along with other criteria such as diversity or steady-state performance. Simulations using parameter ensembles can estimate confidence intervals on model variables, and robustly constrain model predictions, despite having many poorly constrained parameters. In this software note, we present a multiobjective based technique to estimate parameter or models ensembles, the Pareto Optimal Ensemble Technique in the Julia programming language (JuPOETs). JuPOETs integrates simulated annealing with Pareto optimality to estimate ensembles on or near the optimal tradeoff surface between competing training objectives. We demonstrate JuPOETs on a suite of multiobjective problems, including test functions with parameter bounds and system constraints as well as for the identification of a proof-of-concept biochemical model with four conflicting training objectives. JuPOETs identified optimal or near optimal solutions approximately six-fold faster than a corresponding implementation in Octave for the suite of test functions. For the proof-of-concept biochemical model, JuPOETs produced an ensemble of parameters that gave both the mean of the training data for conflicting data sets, while simultaneously estimating parameter sets that performed well on each of the individual objective functions. JuPOETs is a promising approach for the estimation of parameter and model ensembles using multiobjective optimization. JuPOETs can be adapted to solve many problem types, including mixed binary and continuous variable types, bilevel optimization problems and constrained problems without altering the base algorithm. JuPOETs is open
Oil refining in a CO2 constrained world: Effects of carbon pricing on refineries globally
International Nuclear Information System (INIS)
Abdul-Manan, Amir F.N.; Arfaj, Abdullah; Babiker, Hassan
2017-01-01
Six aggregated refinery linear programming (LP) models were developed to represent actual refineries in North America, Latin America, Europe (including the CIS), Middle East, Asia (excluding China) and China. The models were used to conduct regional comparative assessments and to evaluate the effects of carbon pricing on refinery operations globally. We found that the average refinery energy efficiencies for the regions were estimated to range from 92.2% to 95.2%. The well-to-refinery gate carbon intensities for gasoline, diesel and jet fuels were estimated to be 17.1 (16.4–19.4), 13.3 (12.5–14.2) and 10.1 (9.6–10.8) gCO2eq/MJ, respectively. If refineries are forced to at least meet the 2014 regional volume demands for oil products, pricing CO 2 would not have an impact on either refinery productions, efficiency or emissions. If refineries are allowed to re-optimize production slates to reduce CO 2 emissions, refineries would opt to increase gasoline yield at the expense of diesel. This is counter intuitive since gasoline has a higher carbon intensity than diesel. The refinery bias against dieselization creates a supply preference toward a less efficient transportation end use. Here, we argue that if carbon pricing is not administered properly, this can lead to emissions leakage from refineries to the road transport sector. - Highlights: • Investigate actual refinery productions in 6 regions globally. • Refineries already operate at the most efficient levels. • Complex refineries tolerate higher CO 2 prices better. • Carbon pricing induces bias against dieselization. • Identify potential emissions leakage.
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Farouq, S.; Neytcheva, M.
2016-01-01
Roč. 73, č. 3 (2016), s. 631-633 ISSN 1017-1398 R&D Projects: GA MŠk ED1.1.00/02.0070 Institutional support: RVO:68145535 Keywords : PDE-constrained optimization problems * finite elements * iterative solution methods Subject RIV: BA - General Mathematics Impact factor: 1.241, year: 2016 http://link.springer.com/article/10.1007%2Fs11075-016-0111-1
Theory and Algorithms for Global/Local Design Optimization
National Research Council Canada - National Science Library
Haftka, Raphael T
2004-01-01
... the component and overall design as well as on exploration of global optimization algorithms. In the former category, heuristic decomposition was followed with proof that it solves the original problem...
Theory and Algorithms for Global/Local Design Optimization
National Research Council Canada - National Science Library
Watson, Layne T; Guerdal, Zafer; Haftka, Raphael T
2005-01-01
The motivating application for this research is the global/local optimal design of composite aircraft structures such as wings and fuselages, but the theory and algorithms are more widely applicable...
Global optimization of silicon nanowires for efficient parametric processes
DEFF Research Database (Denmark)
Vukovic, Dragana; Xu, Jing; Mørk, Jesper
2013-01-01
We present a global optimization of silicon nanowires for parametric single-pump mixing. For the first time, the effect of surface roughness-induced loss is included in the analysis, significantly influencing the optimum waveguide dimensions.......We present a global optimization of silicon nanowires for parametric single-pump mixing. For the first time, the effect of surface roughness-induced loss is included in the analysis, significantly influencing the optimum waveguide dimensions....
Directory of Open Access Journals (Sweden)
R. Venkata Rao
2016-01-01
Full Text Available The teaching-learning-based optimization (TLBO algorithm is finding a large number of applications in different fields of engineering and science since its introduction in 2011. The major applications are found in electrical engineering, mechanical design, thermal engineering, manufacturing engineering, civil engineering, structural engineering, computer engineering, electronics engineering, physics, chemistry, biotechnology and economics. This paper presents a review of applications of TLBO algorithm and a tutorial for solving the unconstrained and constrained optimization problems. The tutorial is expected to be useful to the beginners.
Directory of Open Access Journals (Sweden)
Arnaut Dierck
2015-01-01
Full Text Available Designing textile antennas for real-life applications requires a design strategy that is able to produce antennas that are optimized over a wide bandwidth for often conflicting characteristics, such as impedance matching, axial ratio, efficiency, and gain, and, moreover, that is able to account for the variations that apply for the characteristics of the unconventional materials used in smart textile systems. In this paper, such a strategy, incorporating a multiobjective constrained Pareto optimization, is presented and applied to the design of a Galileo E6-band antenna with optimal return loss and wide-band axial ratio characteristics. Subsequently, different prototypes of the optimized antenna are fabricated and measured to validate the proposed design strategy.
DEFF Research Database (Denmark)
Liu, Zhaoxi; Wu, Qiuwei; Oren, Shmuel S.
2017-01-01
This paper presents a distribution locational marginal pricing (DLMP) method through chance constrained mixed-integer programming designed to alleviate the possible congestion in the future distribution network with high penetration of electric vehicles (EVs). In order to represent the stochastic...
Zhang, Chenglong; Guo, Ping
2017-10-01
The vague and fuzzy parametric information is a challenging issue in irrigation water management problems. In response to this problem, a generalized fuzzy credibility-constrained linear fractional programming (GFCCFP) model is developed for optimal irrigation water allocation under uncertainty. The model can be derived from integrating generalized fuzzy credibility-constrained programming (GFCCP) into a linear fractional programming (LFP) optimization framework. Therefore, it can solve ratio optimization problems associated with fuzzy parameters, and examine the variation of results under different credibility levels and weight coefficients of possibility and necessary. It has advantages in: (1) balancing the economic and resources objectives directly; (2) analyzing system efficiency; (3) generating more flexible decision solutions by giving different credibility levels and weight coefficients of possibility and (4) supporting in-depth analysis of the interrelationships among system efficiency, credibility level and weight coefficient. The model is applied to a case study of irrigation water allocation in the middle reaches of Heihe River Basin, northwest China. Therefore, optimal irrigation water allocation solutions from the GFCCFP model can be obtained. Moreover, factorial analysis on the two parameters (i.e. λ and γ) indicates that the weight coefficient is a main factor compared with credibility level for system efficiency. These results can be effective for support reasonable irrigation water resources management and agricultural production.
Directory of Open Access Journals (Sweden)
Xing Liu
2014-12-01
Full Text Available Memory and energy optimization strategies are essential for the resource-constrained wireless sensor network (WSN nodes. In this article, a new memory-optimized and energy-optimized multithreaded WSN operating system (OS LiveOS is designed and implemented. Memory cost of LiveOS is optimized by using the stack-shifting hybrid scheduling approach. Different from the traditional multithreaded OS in which thread stacks are allocated statically by the pre-reservation, thread stacks in LiveOS are allocated dynamically by using the stack-shifting technique. As a result, memory waste problems caused by the static pre-reservation can be avoided. In addition to the stack-shifting dynamic allocation approach, the hybrid scheduling mechanism which can decrease both the thread scheduling overhead and the thread stack number is also implemented in LiveOS. With these mechanisms, the stack memory cost of LiveOS can be reduced more than 50% if compared to that of a traditional multithreaded OS. Not is memory cost optimized, but also the energy cost is optimized in LiveOS, and this is achieved by using the multi-core “context aware” and multi-core “power-off/wakeup” energy conservation approaches. By using these approaches, energy cost of LiveOS can be reduced more than 30% when compared to the single-core WSN system. Memory and energy optimization strategies in LiveOS not only prolong the lifetime of WSN nodes, but also make the multithreaded OS feasible to run on the memory-constrained WSN nodes.
Energy Technology Data Exchange (ETDEWEB)
Delbos, F.
2004-11-01
Reflexion tomography allows the determination of a subsurface velocity model from the travel times of seismic waves. The introduction of a priori information in this inverse problem can lead to the resolution of a constrained non-linear least-squares problem. The goal of the thesis is to improve the resolution techniques of this optimization problem, whose main difficulties are its ill-conditioning, its large scale and an expensive cost function in terms of CPU time. Thanks to a detailed study of the problem and to numerous numerical experiments, we justify the use of a sequential quadratic programming method, in which the tangential quadratic programs are solved by an original augmented Lagrangian method. We show the global linear convergence of the latter. The efficiency and robustness of the approach are demonstrated on several synthetic examples and on two real data cases. (author)
Energy Technology Data Exchange (ETDEWEB)
Delbos, F
2004-11-01
Reflexion tomography allows the determination of a subsurface velocity model from the travel times of seismic waves. The introduction of a priori information in this inverse problem can lead to the resolution of a constrained non-linear least-squares problem. The goal of the thesis is to improve the resolution techniques of this optimization problem, whose main difficulties are its ill-conditioning, its large scale and an expensive cost function in terms of CPU time. Thanks to a detailed study of the problem and to numerous numerical experiments, we justify the use of a sequential quadratic programming method, in which the tangential quadratic programs are solved by an original augmented Lagrangian method. We show the global linear convergence of the latter. The efficiency and robustness of the approach are demonstrated on several synthetic examples and on two real data cases. (author)
Directory of Open Access Journals (Sweden)
Chocat Rudy
2015-01-01
Full Text Available The design of complex systems often induces a constrained optimization problem under uncertainty. An adaptation of CMA-ES(λ, μ optimization algorithm is proposed in order to efficiently handle the constraints in the presence of noise. The update mechanisms of the parametrized distribution used to generate the candidate solutions are modified. The constraint handling method allows to reduce the semi-principal axes of the probable research ellipsoid in the directions violating the constraints. The proposed approach is compared to existing approaches on three analytic optimization problems to highlight the efficiency and the robustness of the algorithm. The proposed method is used to design a two stage solid propulsion launch vehicle.
Preconditioners for state-constrained optimal control problems with Moreau-Yosida penalty function
Pearson, John W.; Stoll, Martin; Wathen, Andrew J.
2012-01-01
Optimal control problems with partial differential equations as constraints play an important role in many applications. The inclusion of bound constraints for the state variable poses a significant challenge for optimization methods. Our focus here
Optimizing human activity patterns using global sensitivity analysis.
Fairchild, Geoffrey; Hickmann, Kyle S; Mniszewski, Susan M; Del Valle, Sara Y; Hyman, James M
2014-12-01
Implementing realistic activity patterns for a population is crucial for modeling, for example, disease spread, supply and demand, and disaster response. Using the dynamic activity simulation engine, DASim, we generate schedules for a population that capture regular (e.g., working, eating, and sleeping) and irregular activities (e.g., shopping or going to the doctor). We use the sample entropy (SampEn) statistic to quantify a schedule's regularity for a population. We show how to tune an activity's regularity by adjusting SampEn, thereby making it possible to realistically design activities when creating a schedule. The tuning process sets up a computationally intractable high-dimensional optimization problem. To reduce the computational demand, we use Bayesian Gaussian process regression to compute global sensitivity indices and identify the parameters that have the greatest effect on the variance of SampEn. We use the harmony search (HS) global optimization algorithm to locate global optima. Our results show that HS combined with global sensitivity analysis can efficiently tune the SampEn statistic with few search iterations. We demonstrate how global sensitivity analysis can guide statistical emulation and global optimization algorithms to efficiently tune activities and generate realistic activity patterns. Though our tuning methods are applied to dynamic activity schedule generation, they are general and represent a significant step in the direction of automated tuning and optimization of high-dimensional computer simulations.
Directory of Open Access Journals (Sweden)
Zunaira Nadeem
2018-04-01
Full Text Available In this paper, we design a controller for home energy management based on following meta-heuristic algorithms: teaching learning-based optimization (TLBO, genetic algorithm (GA, firefly algorithm (FA and optimal stopping rule (OSR theory. The principal goal of designing this controller is to reduce the energy consumption of residential sectors while reducing consumer’s electricity bill and maximizing user comfort. Additionally, we propose three hybrid schemes OSR-GA, OSR-TLBO and OSR-FA, by combining the best features of existing algorithms. We have also optimized the desired parameters: peak to average ratio, energy consumption, cost, and user comfort (appliance waiting time for 20, 50, 100 and 200 heterogeneous homes in two steps. In the first step, we obtain the optimal scheduling of home appliances implementing our aforementioned hybrid schemes for single and multiple homes while considering user preferences and threshold base policy. In the second step, we formulate our problem through chance constrained optimization. Simulation results show that proposed hybrid scheduling schemes outperformed for single and multiple homes and they shift the consumer load demand exceeding a predefined threshold to the hours where the electricity price is low thus following the threshold base policy. This helps to reduce electricity cost while considering the comfort of a user by minimizing delay and peak to average ratio. In addition, chance-constrained optimization is used to ensure the scheduling of appliances while considering the uncertainties of a load hence smoothing the load curtailment. The major focus is to keep the appliances power consumption within the power constraint, while keeping power consumption below a pre-defined acceptable level. Moreover, the feasible regions of appliances electricity consumption are calculated which show the relationship between cost and energy consumption and cost and waiting time.
Directory of Open Access Journals (Sweden)
Nebojsa Bacanin
2014-01-01
portfolio model with entropy constraint. Firefly algorithm is one of the latest, very successful swarm intelligence algorithm; however, it exhibits some deficiencies when applied to constrained problems. To overcome lack of exploration power during early iterations, we modified the algorithm and tested it on standard portfolio benchmark data sets used in the literature. Our proposed modified firefly algorithm proved to be better than other state-of-the-art algorithms, while introduction of entropy diversity constraint further improved results.
Computational Approaches to Simulation and Optimization of Global Aircraft Trajectories
Ng, Hok Kwan; Sridhar, Banavar
2016-01-01
This study examines three possible approaches to improving the speed in generating wind-optimal routes for air traffic at the national or global level. They are: (a) using the resources of a supercomputer, (b) running the computations on multiple commercially available computers and (c) implementing those same algorithms into NASAs Future ATM Concepts Evaluation Tool (FACET) and compares those to a standard implementation run on a single CPU. Wind-optimal aircraft trajectories are computed using global air traffic schedules. The run time and wait time on the supercomputer for trajectory optimization using various numbers of CPUs ranging from 80 to 10,240 units are compared with the total computational time for running the same computation on a single desktop computer and on multiple commercially available computers for potential computational enhancement through parallel processing on the computer clusters. This study also re-implements the trajectory optimization algorithm for further reduction of computational time through algorithm modifications and integrates that with FACET to facilitate the use of the new features which calculate time-optimal routes between worldwide airport pairs in a wind field for use with existing FACET applications. The implementations of trajectory optimization algorithms use MATLAB, Python, and Java programming languages. The performance evaluations are done by comparing their computational efficiencies and based on the potential application of optimized trajectories. The paper shows that in the absence of special privileges on a supercomputer, a cluster of commercially available computers provides a feasible approach for national and global air traffic system studies.
Energy Technology Data Exchange (ETDEWEB)
Wei, J [City College of New York, New York, NY (United States); Chao, M [The Mount Sinai Medical Center, New York, NY (United States)
2016-06-15
Purpose: To develop a novel strategy to extract the respiratory motion of the thoracic diaphragm from kilovoltage cone beam computed tomography (CBCT) projections by a constrained linear regression optimization technique. Methods: A parabolic function was identified as the geometric model and was employed to fit the shape of the diaphragm on the CBCT projections. The search was initialized by five manually placed seeds on a pre-selected projection image. Temporal redundancies, the enabling phenomenology in video compression and encoding techniques, inherent in the dynamic properties of the diaphragm motion together with the geometrical shape of the diaphragm boundary and the associated algebraic constraint that significantly reduced the searching space of viable parabolic parameters was integrated, which can be effectively optimized by a constrained linear regression approach on the subsequent projections. The innovative algebraic constraints stipulating the kinetic range of the motion and the spatial constraint preventing any unphysical deviations was able to obtain the optimal contour of the diaphragm with minimal initialization. The algorithm was assessed by a fluoroscopic movie acquired at anteriorposterior fixed direction and kilovoltage CBCT projection image sets from four lung and two liver patients. The automatic tracing by the proposed algorithm and manual tracking by a human operator were compared in both space and frequency domains. Results: The error between the estimated and manual detections for the fluoroscopic movie was 0.54mm with standard deviation (SD) of 0.45mm, while the average error for the CBCT projections was 0.79mm with SD of 0.64mm for all enrolled patients. The submillimeter accuracy outcome exhibits the promise of the proposed constrained linear regression approach to track the diaphragm motion on rotational projection images. Conclusion: The new algorithm will provide a potential solution to rendering diaphragm motion and ultimately
International Nuclear Information System (INIS)
Wei, J; Chao, M
2016-01-01
Purpose: To develop a novel strategy to extract the respiratory motion of the thoracic diaphragm from kilovoltage cone beam computed tomography (CBCT) projections by a constrained linear regression optimization technique. Methods: A parabolic function was identified as the geometric model and was employed to fit the shape of the diaphragm on the CBCT projections. The search was initialized by five manually placed seeds on a pre-selected projection image. Temporal redundancies, the enabling phenomenology in video compression and encoding techniques, inherent in the dynamic properties of the diaphragm motion together with the geometrical shape of the diaphragm boundary and the associated algebraic constraint that significantly reduced the searching space of viable parabolic parameters was integrated, which can be effectively optimized by a constrained linear regression approach on the subsequent projections. The innovative algebraic constraints stipulating the kinetic range of the motion and the spatial constraint preventing any unphysical deviations was able to obtain the optimal contour of the diaphragm with minimal initialization. The algorithm was assessed by a fluoroscopic movie acquired at anteriorposterior fixed direction and kilovoltage CBCT projection image sets from four lung and two liver patients. The automatic tracing by the proposed algorithm and manual tracking by a human operator were compared in both space and frequency domains. Results: The error between the estimated and manual detections for the fluoroscopic movie was 0.54mm with standard deviation (SD) of 0.45mm, while the average error for the CBCT projections was 0.79mm with SD of 0.64mm for all enrolled patients. The submillimeter accuracy outcome exhibits the promise of the proposed constrained linear regression approach to track the diaphragm motion on rotational projection images. Conclusion: The new algorithm will provide a potential solution to rendering diaphragm motion and ultimately
Deterministic global optimization an introduction to the diagonal approach
Sergeyev, Yaroslav D
2017-01-01
This book begins with a concentrated introduction into deterministic global optimization and moves forward to present new original results from the authors who are well known experts in the field. Multiextremal continuous problems that have an unknown structure with Lipschitz objective functions and functions having the first Lipschitz derivatives defined over hyperintervals are examined. A class of algorithms using several Lipschitz constants is introduced which has its origins in the DIRECT (DIviding RECTangles) method. This new class is based on an efficient strategy that is applied for the search domain partitioning. In addition a survey on derivative free methods and methods using the first derivatives is given for both one-dimensional and multi-dimensional cases. Non-smooth and smooth minorants and acceleration techniques that can speed up several classes of global optimization methods with examples of applications and problems arising in numerical testing of global optimization algorithms are discussed...
Application of surrogate-based global optimization to aerodynamic design
Pérez, Esther
2016-01-01
Aerodynamic design, like many other engineering applications, is increasingly relying on computational power. The growing need for multi-disciplinarity and high fidelity in design optimization for industrial applications requires a huge number of repeated simulations in order to find an optimal design candidate. The main drawback is that each simulation can be computationally expensive – this becomes an even bigger issue when used within parametric studies, automated search or optimization loops, which typically may require thousands of analysis evaluations. The core issue of a design-optimization problem is the search process involved. However, when facing complex problems, the high-dimensionality of the design space and the high-multi-modality of the target functions cannot be tackled with standard techniques. In recent years, global optimization using meta-models has been widely applied to design exploration in order to rapidly investigate the design space and find sub-optimal solutions. Indeed, surrogat...
Global Optimization for Bus Line Timetable Setting Problem
Directory of Open Access Journals (Sweden)
Qun Chen
2014-01-01
Full Text Available This paper defines bus timetables setting problem during each time period divided in terms of passenger flow intensity; it is supposed that passengers evenly arrive and bus runs are set evenly; the problem is to determine bus runs assignment in each time period to minimize the total waiting time of passengers on platforms if the number of the total runs is known. For such a multistage decision problem, this paper designed a dynamic programming algorithm to solve it. Global optimization procedures using dynamic programming are developed. A numerical example about bus runs assignment optimization of a single line is given to demonstrate the efficiency of the proposed methodology, showing that optimizing buses’ departure time using dynamic programming can save computational time and find the global optimal solution.
A dynamic global and local combined particle swarm optimization algorithm
International Nuclear Information System (INIS)
Jiao Bin; Lian Zhigang; Chen Qunxian
2009-01-01
Particle swarm optimization (PSO) algorithm has been developing rapidly and many results have been reported. PSO algorithm has shown some important advantages by providing high speed of convergence in specific problems, but it has a tendency to get stuck in a near optimal solution and one may find it difficult to improve solution accuracy by fine tuning. This paper presents a dynamic global and local combined particle swarm optimization (DGLCPSO) algorithm to improve the performance of original PSO, in which all particles dynamically share the best information of the local particle, global particle and group particles. It is tested with a set of eight benchmark functions with different dimensions and compared with original PSO. Experimental results indicate that the DGLCPSO algorithm improves the search performance on the benchmark functions significantly, and shows the effectiveness of the algorithm to solve optimization problems.
Constrained Optimal Stochastic Control of Non-Linear Wave Energy Point Absorbers
DEFF Research Database (Denmark)
Sichani, Mahdi Teimouri; Chen, Jian-Bing; Kramer, Morten
2014-01-01
to extract energy. Constrains are enforced on the control force to prevent large structural stresses in the floater at specific hot spots with the risk of inducing fatigue damage, or because the demanded control force cannot be supplied by the actuator system due to saturation. Further, constraints...... are enforced on the motion of the floater to prevent it from hitting the bottom of the sea or to make unacceptable jumps out of the water. The applied control law, which is of the feedback type with feedback from the displacement, velocity, and acceleration of the floater, contains two unprovided gain...
Robust optimization methods for chance constrained, simulation-based, and bilevel problems
Yanikoglu, I.
2014-01-01
The objective of robust optimization is to find solutions that are immune to the uncertainty of the parameters in a mathematical optimization problem. It requires that the constraints of a given problem should be satisfied for all realizations of the uncertain parameters in a so-called uncertainty
Trade-offs and efficiencies in optimal budget-constrained multispecies corridor networks
Bistra Dilkina; Rachel Houtman; Carla P. Gomes; Claire A. Montgomery; Kevin S. McKelvey; Katherine Kendall; Tabitha A. Graves; Richard Bernstein; Michael K. Schwartz
2016-01-01
Conservation biologists recognize that a system of isolated protected areas will be necessary but insufficient to meet biodiversity objectives. Current approaches to connecting core conservation areas through corridors consider optimal corridor placement based on a single optimization goal: commonly, maximizing the movement for a target species across a...
Optimization of constrained multiple-objective reliability problems using evolutionary algorithms
International Nuclear Information System (INIS)
Salazar, Daniel; Rocco, Claudio M.; Galvan, Blas J.
2006-01-01
This paper illustrates the use of multi-objective optimization to solve three types of reliability optimization problems: to find the optimal number of redundant components, find the reliability of components, and determine both their redundancy and reliability. In general, these problems have been formulated as single objective mixed-integer non-linear programming problems with one or several constraints and solved by using mathematical programming techniques or special heuristics. In this work, these problems are reformulated as multiple-objective problems (MOP) and then solved by using a second-generation Multiple-Objective Evolutionary Algorithm (MOEA) that allows handling constraints. The MOEA used in this paper (NSGA-II) demonstrates the ability to identify a set of optimal solutions (Pareto front), which provides the Decision Maker with a complete picture of the optimal solution space. Finally, the advantages of both MOP and MOEA approaches are illustrated by solving four redundancy problems taken from the literature
Optimization of constrained multiple-objective reliability problems using evolutionary algorithms
Energy Technology Data Exchange (ETDEWEB)
Salazar, Daniel [Instituto de Sistemas Inteligentes y Aplicaciones Numericas en Ingenieria (IUSIANI), Division de Computacion Evolutiva y Aplicaciones (CEANI), Universidad de Las Palmas de Gran Canaria, Islas Canarias (Spain) and Facultad de Ingenieria, Universidad Central Venezuela, Caracas (Venezuela)]. E-mail: danielsalazaraponte@gmail.com; Rocco, Claudio M. [Facultad de Ingenieria, Universidad Central Venezuela, Caracas (Venezuela)]. E-mail: crocco@reacciun.ve; Galvan, Blas J. [Instituto de Sistemas Inteligentes y Aplicaciones Numericas en Ingenieria (IUSIANI), Division de Computacion Evolutiva y Aplicaciones (CEANI), Universidad de Las Palmas de Gran Canaria, Islas Canarias (Spain)]. E-mail: bgalvan@step.es
2006-09-15
This paper illustrates the use of multi-objective optimization to solve three types of reliability optimization problems: to find the optimal number of redundant components, find the reliability of components, and determine both their redundancy and reliability. In general, these problems have been formulated as single objective mixed-integer non-linear programming problems with one or several constraints and solved by using mathematical programming techniques or special heuristics. In this work, these problems are reformulated as multiple-objective problems (MOP) and then solved by using a second-generation Multiple-Objective Evolutionary Algorithm (MOEA) that allows handling constraints. The MOEA used in this paper (NSGA-II) demonstrates the ability to identify a set of optimal solutions (Pareto front), which provides the Decision Maker with a complete picture of the optimal solution space. Finally, the advantages of both MOP and MOEA approaches are illustrated by solving four redundancy problems taken from the literature.
Crown, William; Buyukkaramikli, Nasuh; Thokala, Praveen; Morton, Alec; Sir, Mustafa Y; Marshall, Deborah A; Tosh, Jon; Padula, William V; Ijzerman, Maarten J; Wong, Peter K; Pasupathy, Kalyan S
2017-03-01
Providing health services with the greatest possible value to patients and society given the constraints imposed by patient characteristics, health care system characteristics, budgets, and so forth relies heavily on the design of structures and processes. Such problems are complex and require a rigorous and systematic approach to identify the best solution. Constrained optimization is a set of methods designed to identify efficiently and systematically the best solution (the optimal solution) to a problem characterized by a number of potential solutions in the presence of identified constraints. This report identifies 1) key concepts and the main steps in building an optimization model; 2) the types of problems for which optimal solutions can be determined in real-world health applications; and 3) the appropriate optimization methods for these problems. We first present a simple graphical model based on the treatment of "regular" and "severe" patients, which maximizes the overall health benefit subject to time and budget constraints. We then relate it back to how optimization is relevant in health services research for addressing present day challenges. We also explain how these mathematical optimization methods relate to simulation methods, to standard health economic analysis techniques, and to the emergent fields of analytics and machine learning. Copyright © 2017 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.
Bouaziz, Laurène; Hegnauer, Mark; Schellekens, Jaap; Sperna Weiland, Frederiek; ten Velden, Corine
2017-04-01
In many countries, data is scarce, incomplete and often not easily shared. In these cases, global satellite and reanalysis data provide an alternative to assess water resources. To assess water resources in Azerbaijan, a completely distributed and physically based hydrological wflow-sbm model was set-up for the entire Kura basin. We used SRTM elevation data, a locally available river map and one from OpenStreetMap to derive the drainage direction network at the model resolution of approximately 1x1 km. OpenStreetMap data was also used to derive the fraction of paved area per cell to account for the reduced infiltration capacity (c.f. Schellekens et al. 2014). We used the results of a global study to derive root zone capacity based on climate data (Wang-Erlandsson et al., 2016). To account for the variation in vegetation cover over the year, monthly averages of Leaf Area Index, based on MODIS data, were used. For the soil-related parameters, we used global estimates as provided by Dai et al. (2013). This enabled the rapid derivation of a first estimate of parameter values for our hydrological model. Digitized local meteorological observations were scarce and available only for limited time period. Therefore several sources of global meteorological data were evaluated: (1) EU-WATCH global precipitation, temperature and derived potential evaporation for the period 1958-2001 (Harding et al., 2011), (2) WFDEI precipitation, temperature and derived potential evaporation for the period 1979-2014 (by Weedon et al., 2014), (3) MSWEP precipitation (Beck et al., 2016) and (4) local precipitation data from more than 200 stations in the Kura basin were available from the NOAA website for a period up to 1991. The latter, together with data archives from Azerbaijan, were used as a benchmark to evaluate the global precipitation datasets for the overlapping period 1958-1991. By comparing the datasets, we found that monthly mean precipitation of EU-WATCH and WFDEI coincided well
Wang, Y.; Stek, P. C.; Su, H.; Jiang, J. H.; Livesey, N. J.; Santee, M. L.
2014-12-01
Over the past century, global average surface temperature has warmed by about 0.16°C/decade, largely due to anthropogenic increases in well-mixed greenhouse gases. However, the trend in global surface temperatures has been nearly flat since 2000, raising a question regarding the exploration of the drivers of climate change. Water vapor is a strong greenhouse gas in the atmosphere. Previous studies suggested that the sudden decrease of stratospheric water vapor (SWV) around 2000 may have contributed to the stall of global warming. Since 2004, the SWV observed by Microwave Limb Sounder (MLS) on Aura satellite has shown a slow recovery. The role of recent SWV variations in global warming has not been quantified. We employ a coupled atmosphere-ocean climate model, the NCAR CESM, to address this issue. It is found that the CESM underestimates the stratospheric water vapor by about 1 ppmv due to limited representations of the stratospheric dynamic and chemical processes important for water vapor variabilities. By nudging the modeled SWV to the MLS observation, we find that increasing SWV by 1 ppmv produces a robust surface warming about 0.2°C in global-mean when the model reaches equilibrium. Conversely, the sudden drop of SWV from 2000 to 2004 would cause a surface cooling about -0.08°C in global-mean. On the other hand, imposing the observed linear trend of SWV based on the 10-year observation of MLS in the CESM yields a rather slow surface warming, about 0.04°C/decade. Our model experiments suggest that SWV contributes positively to the global surface temperature variation, although it may not be the dominant factor that drives the recent global warming hiatus. Additional sensitivity experiments show that the impact of SWV on surface climate is mostly governed by the SWV amount at 100 hPa in the tropics. Furthermore, the atmospheric model simulations driven by observed sea surface temperature (SST) show that the inter-annual variation of SWV follows that of SST
Global-local optimization of flapping kinematics in hovering flight
Ghommem, Mehdi; Hajj, M. R.; Mook, Dean T.; Stanford, Bret K.; Bé ran, Philip S.; Watson, Layne T.
2013-01-01
The kinematics of a hovering wing are optimized by combining the 2-d unsteady vortex lattice method with a hybrid of global and local optimization algorithms. The objective is to minimize the required aerodynamic power under a lift constraint. The hybrid optimization is used to efficiently navigate the complex design space due to wing-wake interference present in hovering aerodynamics. The flapping wing is chosen so that its chord length and flapping frequency match the morphological and flight properties of two insects with different masses. The results suggest that imposing a delay between the different oscillatory motions defining the flapping kinematics, and controlling the way through which the wing rotates at the end of each half stroke can improve aerodynamic power under a lift constraint. Furthermore, our optimization analysis identified optimal kinematics that agree fairly well with observed insect kinematics, as well as previously published numerical results.
Global-local optimization of flapping kinematics in hovering flight
Ghommem, Mehdi
2013-06-01
The kinematics of a hovering wing are optimized by combining the 2-d unsteady vortex lattice method with a hybrid of global and local optimization algorithms. The objective is to minimize the required aerodynamic power under a lift constraint. The hybrid optimization is used to efficiently navigate the complex design space due to wing-wake interference present in hovering aerodynamics. The flapping wing is chosen so that its chord length and flapping frequency match the morphological and flight properties of two insects with different masses. The results suggest that imposing a delay between the different oscillatory motions defining the flapping kinematics, and controlling the way through which the wing rotates at the end of each half stroke can improve aerodynamic power under a lift constraint. Furthermore, our optimization analysis identified optimal kinematics that agree fairly well with observed insect kinematics, as well as previously published numerical results.
Dispositional Optimism and Terminal Decline in Global Quality of Life
Zaslavsky, Oleg; Palgi, Yuval; Rillamas-Sun, Eileen; LaCroix, Andrea Z.; Schnall, Eliezer; Woods, Nancy F.; Cochrane, Barbara B.; Garcia, Lorena; Hingle, Melanie; Post, Stephen; Seguin, Rebecca; Tindle, Hilary; Shrira, Amit
2015-01-01
We examined whether dispositional optimism relates to change in global quality of life (QOL) as a function of either chronological age or years to impending death. We used a sample of 2,096 deceased postmenopausal women from the Women's Health Initiative clinical trials who were enrolled in the 2005-2010 Extension Study and for whom at least 1…
Keeping global warming within 1.5 °C constrains emergence of aridification
Park, Chang-Eui; Jeong, Su-Jong; Joshi, Manoj; Osborn, Timothy J.; Ho, Chang-Hoi; Piao, Shilong; Chen, Deliang; Liu, Junguo; Yang, Hong; Park, Hoonyoung; Kim, Baek-Min; Feng, Song
2018-01-01
Aridity—the ratio of atmospheric water supply (precipitation; P) to demand (potential evapotranspiration; PET)—is projected to decrease (that is, areas will become drier) as a consequence of anthropogenic climate change, exacerbating land degradation and desertification1-6. However, the timing of significant aridification relative to natural variability—defined here as the time of emergence for aridification (ToEA)—is unknown, despite its importance in designing and implementing mitigation policies7-10. Here we estimate ToEA from projections of 27 global climate models (GCMs) under representative concentration pathways (RCPs) RCP4.5 and RCP8.5, and in doing so, identify where emergence occurs before global mean warming reaches 1.5 °C and 2 °C above the pre-industrial level. On the basis of the ensemble median ToEA for each grid cell, aridification emerges over 32% (RCP4.5) and 24% (RCP8.5) of the total land surface before the ensemble median of global mean temperature change reaches 2 °C in each scenario. Moreover, ToEA is avoided in about two-thirds of the above regions if the maximum global warming level is limited to 1.5 °C. Early action for accomplishing the 1.5 °C temperature goal can therefore markedly reduce the likelihood that large regions will face substantial aridification and related impacts.
Houweling, S; Dentener, F; Lelieveld, J
2000-01-01
Previous attempts to quantify the global source strength of CH4 from natural wetlands have resulted in a range of 90-260 TE(CH4) yr(-1). This relatively uncertain estimate significantly limits our understanding of atmospheric methane. In this study we reduce this uncertainty by simulating
Global Optimization of Nonlinear Blend-Scheduling Problems
Directory of Open Access Journals (Sweden)
Pedro A. Castillo Castillo
2017-04-01
Full Text Available The scheduling of gasoline-blending operations is an important problem in the oil refining industry. This problem not only exhibits the combinatorial nature that is intrinsic to scheduling problems, but also non-convex nonlinear behavior, due to the blending of various materials with different quality properties. In this work, a global optimization algorithm is proposed to solve a previously published continuous-time mixed-integer nonlinear scheduling model for gasoline blending. The model includes blend recipe optimization, the distribution problem, and several important operational features and constraints. The algorithm employs piecewise McCormick relaxation (PMCR and normalized multiparametric disaggregation technique (NMDT to compute estimates of the global optimum. These techniques partition the domain of one of the variables in a bilinear term and generate convex relaxations for each partition. By increasing the number of partitions and reducing the domain of the variables, the algorithm is able to refine the estimates of the global solution. The algorithm is compared to two commercial global solvers and two heuristic methods by solving four examples from the literature. Results show that the proposed global optimization algorithm performs on par with commercial solvers but is not as fast as heuristic approaches.
Directory of Open Access Journals (Sweden)
A. P. Karpenko
2014-01-01
Full Text Available We consider a class of stochastic search algorithms of global optimization which in various publications are called behavioural, intellectual, metaheuristic, inspired by the nature, swarm, multi-agent, population, etc. We use the last term.Experience in using the population algorithms to solve challenges of global optimization shows that application of one such algorithm may not always effective. Therefore now great attention is paid to hybridization of population algorithms of global optimization. Hybrid algorithms unite various algorithms or identical algorithms, but with various values of free parameters. Thus efficiency of one algorithm can compensate weakness of another.The purposes of the work are development of hybrid algorithm of global optimization based on known algorithms of harmony search (HS and swarm of particles (PSO, software implementation of algorithm, study of its efficiency using a number of known benchmark problems, and a problem of dimensional optimization of truss structure.We set a problem of global optimization, consider basic algorithms of HS and PSO, give a flow chart of the offered hybrid algorithm called PSO HS , present results of computing experiments with developed algorithm and software, formulate main results of work and prospects of its development.
Multi-Objective Design Optimization of an Over-Constrained Flexure-Based Amplifier
Directory of Open Access Journals (Sweden)
Yuan Ni
2015-07-01
Full Text Available The optimizing design for enhancement of the micro performance of manipulator based on analytical models is investigated in this paper. By utilizing the established uncanonical linear homogeneous equations, the quasi-static analytical model of the micro-manipulator is built, and the theoretical calculation results are tested by FEA simulations. To provide a theoretical basis for a micro-manipulator being used in high-precision engineering applications, this paper investigates the modal property based on the analytical model. Based on the finite element method, with multipoint constraint equations, the model is built and the results have a good match with the simulation. The following parametric influences studied show that the influences of other objectives on one objective are complicated. Consequently, the multi-objective optimization by the derived analytical models is carried out to find out the optimal solutions of the manipulator. Besides the inner relationships among these design objectives during the optimization process are discussed.
Orito, Yukiko; Yamamoto, Hisashi; Tsujimura, Yasuhiro; Kambayashi, Yasushi
The portfolio optimizations are to determine the proportion-weighted combination in the portfolio in order to achieve investment targets. This optimization is one of the multi-dimensional combinatorial optimizations and it is difficult for the portfolio constructed in the past period to keep its performance in the future period. In order to keep the good performances of portfolios, we propose the extended information ratio as an objective function, using the information ratio, beta, prime beta, or correlation coefficient in this paper. We apply the simulated annealing (SA) to optimize the portfolio employing the proposed ratio. For the SA, we make the neighbor by the operation that changes the structure of the weights in the portfolio. In the numerical experiments, we show that our portfolios keep the good performances when the market trend of the future period becomes different from that of the past period.
Automatic analog IC sizing and optimization constrained with PVT corners and layout effects
Lourenço, Nuno; Horta, Nuno
2017-01-01
This book introduces readers to a variety of tools for automatic analog integrated circuit (IC) sizing and optimization. The authors provide a historical perspective on the early methods proposed to tackle automatic analog circuit sizing, with emphasis on the methodologies to size and optimize the circuit, and on the methodologies to estimate the circuit’s performance. The discussion also includes robust circuit design and optimization and the most recent advances in layout-aware analog sizing approaches. The authors describe a methodology for an automatic flow for analog IC design, including details of the inputs and interfaces, multi-objective optimization techniques, and the enhancements made in the base implementation by using machine leaning techniques. The Gradient model is discussed in detail, along with the methods to include layout effects in the circuit sizing. The concepts and algorithms of all the modules are thoroughly described, enabling readers to reproduce the methodologies, improve the qual...
A non-penalty recurrent neural network for solving a class of constrained optimization problems.
Hosseini, Alireza
2016-01-01
In this paper, we explain a methodology to analyze convergence of some differential inclusion-based neural networks for solving nonsmooth optimization problems. For a general differential inclusion, we show that if its right hand-side set valued map satisfies some conditions, then solution trajectory of the differential inclusion converges to optimal solution set of its corresponding in optimization problem. Based on the obtained methodology, we introduce a new recurrent neural network for solving nonsmooth optimization problems. Objective function does not need to be convex on R(n) nor does the new neural network model require any penalty parameter. We compare our new method with some penalty-based and non-penalty based models. Moreover for differentiable cases, we implement circuit diagram of the new neural network. Copyright © 2015 Elsevier Ltd. All rights reserved.
A one-layer recurrent neural network for constrained nonconvex optimization.
Li, Guocheng; Yan, Zheng; Wang, Jun
2015-01-01
In this paper, a one-layer recurrent neural network is proposed for solving nonconvex optimization problems subject to general inequality constraints, designed based on an exact penalty function method. It is proved herein that any neuron state of the proposed neural network is convergent to the feasible region in finite time and stays there thereafter, provided that the penalty parameter is sufficiently large. The lower bounds of the penalty parameter and convergence time are also estimated. In addition, any neural state of the proposed neural network is convergent to its equilibrium point set which satisfies the Karush-Kuhn-Tucker conditions of the optimization problem. Moreover, the equilibrium point set is equivalent to the optimal solution to the nonconvex optimization problem if the objective function and constraints satisfy given conditions. Four numerical examples are provided to illustrate the performances of the proposed neural network.
Optimization of the box-girder of overhead crane with constrained ...
African Journals Online (AJOL)
haroun
Keywords: Overhead crane - Box-girder - New bat algorithm - level of ... much more efficiency and robustness compared to the genetic algorithm (GA) and PSO ...... optimization: developments, applications and resources," in Evolutionary.
Wang, Lingfeng; Singh, Chanan
2007-01-01
Source: Swarm Intelligence: Focus on Ant and Particle Swarm Optimization, Book edited by: Felix T. S. Chan and Manoj Kumar Tiwari, ISBN 978-3-902613-09-7, pp. 532, December 2007, Itech Education and Publishing, Vienna, Austria
Direct Speed Control of PMSM Drive Using SDRE and Convex Constrained Optimization
Czech Academy of Sciences Publication Activity Database
Šmídl, V.; Janouš, Š.; Adam, Lukáš; Peroutka, Z.
2018-01-01
Roč. 65, č. 1 (2018), s. 532-542 ISSN 1932-4529 Grant - others:GA MŠk(CZ) LO1607 Institutional support: RVO:67985556 Keywords : Velocity control * Optimization * Stators * Voltage control * Predictive control * Optimal control * Rotors Subject RIV: BD - Theory of Information Impact factor: 10.710, year: 2016 http://library.utia.cas.cz/separaty/2017/AS/smidl-0481225.pdf
DEFF Research Database (Denmark)
Sadegh, Payman
1997-01-01
This paper deals with a projection algorithm for stochastic approximation using simultaneous perturbation gradient approximation for optimization under inequality constraints where no direct gradient of the loss function is available and the inequality constraints are given as explicit functions...... of the optimization parameters. It is shown that, under application of the projection algorithm, the parameter iterate converges almost surely to a Kuhn-Tucker point, The procedure is illustrated by a numerical example, (C) 1997 Elsevier Science Ltd....
Global Sufficient Optimality Conditions for a Special Cubic Minimization Problem
Directory of Open Access Journals (Sweden)
Xiaomei Zhang
2012-01-01
Full Text Available We present some sufficient global optimality conditions for a special cubic minimization problem with box constraints or binary constraints by extending the global subdifferential approach proposed by V. Jeyakumar et al. (2006. The present conditions generalize the results developed in the work of V. Jeyakumar et al. where a quadratic minimization problem with box constraints or binary constraints was considered. In addition, a special diagonal matrix is constructed, which is used to provide a convenient method for justifying the proposed sufficient conditions. Then, the reformulation of the sufficient conditions follows. It is worth noting that this reformulation is also applicable to the quadratic minimization problem with box or binary constraints considered in the works of V. Jeyakumar et al. (2006 and Y. Wang et al. (2010. Finally some examples demonstrate that our optimality conditions can effectively be used for identifying global minimizers of the certain nonconvex cubic minimization problem.
Optimal Coordinated EV Charging with Reactive Power Support in Constrained Distribution Grids
Energy Technology Data Exchange (ETDEWEB)
Paudyal, Sumit; Ceylan, Oğuzhan; Bhattarai, Bishnu P.; Myers, Kurt S.
2017-07-01
Electric vehicle (EV) charging/discharging can take place in any P-Q quadrants, which means EVs could support reactive power to the grid while charging the battery. In controlled charging schemes, distribution system operator (DSO) coordinates with the charging of EV fleets to ensure grid’s operating constraints are not violated. In fact, this refers to DSO setting upper bounds on power limits for EV charging. In this work, we demonstrate that if EVs inject reactive power into the grid while charging, DSO could issue higher upper bounds on the active power limits for the EVs for the same set of grid constraints. We demonstrate the concept in an 33-node test feeder with 1,500 EVs. Case studies show that in constrained distribution grids in coordinated charging, average costs of EV charging could be reduced if the charging takes place in the fourth P-Q quadrant compared to charging with unity power factor.
DEFF Research Database (Denmark)
Øjelund, Henrik; Sadegh, Payman
2000-01-01
be obtained. This paper presents a new approach for system modelling under partial (global) information (or the so called Gray-box modelling) that seeks to perserve the benefits of the global as well as local methodologies sithin a unified framework. While the proposed technique relies on local approximations......Local function approximations concern fitting low order models to weighted data in neighbourhoods of the points where the approximations are desired. Despite their generality and convenience of use, local models typically suffer, among others, from difficulties arising in physical interpretation...... simultaneously with the (local estimates of) function values. The approach is applied to modelling of a linear time variant dynamic system under prior linear time invariant structure where local regression fails as a result of high dimensionality....
Liang, Qing; Newman, Paul A.; Daniel, John S.; Reimann, Stefan; Hall, Bradley D.; Dutton, Geoff; Kuijpers, Lambert J. M.
2014-07-01
Carbon tetrachloride (CCl4) is a major anthropogenic ozone-depleting substance and greenhouse gas and has been regulated under the Montreal Protocol. However, the near-zero 2007-2012 emissions estimate based on the UNEP reported production and feedstock usage cannot be reconciled with the observed slow decline of atmospheric concentrations and the inter-hemispheric gradient (IHG) for CCl4. Our 3-D model simulations suggest that the observed IHG (1.5 ± 0.2 ppt for 2000-2012) is primarily caused by ongoing current emissions, while ocean and soil losses and stratosphere-troposphere exchange together contribute a small negative gradient (~0 - -0.3 ppt). Using the observed CCl4 global trend and IHG, we deduce that the mean global emissions for the 2000-2012 period are 393445 Gg/yr (~30% of the peak 1980s emissions) and a corresponding total lifetime of 353732 years.
Khan, M. M. A.; Romoli, L.; Fiaschi, M.; Dini, G.; Sarri, F.
2011-02-01
This paper presents an experimental design approach to process parameter optimization for the laser welding of martensitic AISI 416 and AISI 440FSe stainless steels in a constrained overlap configuration in which outer shell was 0.55 mm thick. To determine the optimal laser-welding parameters, a set of mathematical models were developed relating welding parameters to each of the weld characteristics. These were validated both statistically and experimentally. The quality criteria set for the weld to determine optimal parameters were the minimization of weld width and the maximization of weld penetration depth, resistance length and shearing force. Laser power and welding speed in the range 855-930 W and 4.50-4.65 m/min, respectively, with a fiber diameter of 300 μm were identified as the optimal set of process parameters. However, the laser power and welding speed can be reduced to 800-840 W and increased to 4.75-5.37 m/min, respectively, to obtain stronger and better welds.
Directory of Open Access Journals (Sweden)
Yakai Xu
2017-01-01
Full Text Available Dynamic stiffness and damping of the headstock, which is a critical component of precision horizontal machining center, are two main factors that influence machining accuracy and surface finish quality. Constrained Layer Damping (CLD structure is proved to be effective in raising damping capacity for the thin plate and shell structures. In this paper, one kind of high damping material is utilized on the headstock to improve damping capacity. The dynamic characteristic of the hybrid headstock is investigated analytically and experimentally. The results demonstrate that the resonant response amplitudes of the headstock with damping material can decrease significantly compared to original cast structure. To obtain the optimal configuration of damping material, a topology optimization method based on the Evolutionary Structural Optimization (ESO is implemented. Modal Strain Energy (MSE method is employed to analyze the damping and to derive the sensitivity of the modal loss factor. The optimization results indicate that the added weight of damping material decreases by 50%; meanwhile the first two orders of modal loss factor decrease by less than 23.5% compared to the original structure.
International Nuclear Information System (INIS)
Sankaran, Sethuraman; Audet, Charles; Marsden, Alison L.
2010-01-01
Recent advances in coupling novel optimization methods to large-scale computing problems have opened the door to tackling a diverse set of physically realistic engineering design problems. A large computational overhead is associated with computing the cost function for most practical problems involving complex physical phenomena. Such problems are also plagued with uncertainties in a diverse set of parameters. We present a novel stochastic derivative-free optimization approach for tackling such problems. Our method extends the previously developed surrogate management framework (SMF) to allow for uncertainties in both simulation parameters and design variables. The stochastic collocation scheme is employed for stochastic variables whereas Kriging based surrogate functions are employed for the cost function. This approach is tested on four numerical optimization problems and is shown to have significant improvement in efficiency over traditional Monte-Carlo schemes. Problems with multiple probabilistic constraints are also discussed.
Battaglia, G.; Joos, F.
2018-01-01
Nitrous oxide (N2O) is a potent greenhouse gas (GHG) and ozone destructing agent; yet global estimates of N2O emissions are uncertain. Marine N2O stems from nitrification and denitrification processes which depend on organic matter cycling and dissolved oxygen (O2). We introduce N2O as an obligate intermediate product of denitrification and as an O2-dependent by-product from nitrification in the Bern3D ocean model. A large model ensemble is used to probabilistically constrain modern and to project marine N2O production for a low (Representative Concentration Pathway (RCP)2.6) and high GHG (RCP8.5) scenario extended to A.D. 10,000. Water column N2O and surface ocean partial pressure N2O data serve as constraints in this Bayesian framework. The constrained median for modern N2O production is 4.5 (±1σ range: 3.0 to 6.1) Tg N yr-1, where 4.5% stems from denitrification. Modeled denitrification is 65.1 (40.9 to 91.6) Tg N yr-1, well within current estimates. For high GHG forcing, N2O production decreases by 7.7% over this century due to decreasing organic matter export and remineralization. Thereafter, production increases slowly by 21% due to widespread deoxygenation and high remineralization. Deoxygenation peaks in two millennia, and the global O2 inventory is reduced by a factor of 2 compared to today. Net denitrification is responsible for 7.8% of the long-term increase in N2O production. On millennial timescales, marine N2O emissions constitute a small, positive feedback to climate change. Our simulations reveal tight coupling between the marine carbon cycle, O2, N2O, and climate.
Design Optimization of Time- and Cost-Constrained Fault-Tolerant Distributed Embedded Systems
DEFF Research Database (Denmark)
Izosimov, Viacheslav; Pop, Paul; Eles, Petru
2005-01-01
In this paper we present an approach to the design optimization of fault-tolerant embedded systems for safety-critical applications. Processes are statically scheduled and communications are performed using the time-triggered protocol. We use process re-execution and replication for tolerating...... transient faults. Our design optimization approach decides the mapping of processes to processors and the assignment of fault-tolerant policies to processes such that transient faults are tolerated and the timing constraints of the application are satisfied. We present several heuristics which are able...
Preconditioners for state-constrained optimal control problems with Moreau-Yosida penalty function
Pearson, John W.
2012-11-21
Optimal control problems with partial differential equations as constraints play an important role in many applications. The inclusion of bound constraints for the state variable poses a significant challenge for optimization methods. Our focus here is on the incorporation of the constraints via the Moreau-Yosida regularization technique. This method has been studied recently and has proven to be advantageous compared with other approaches. In this paper, we develop robust preconditioners for the efficient solution of the Newton steps associated with the fast solution of the Moreau-Yosida regularized problem. Numerical results illustrate the efficiency of our approach. © 2012 John Wiley & Sons, Ltd.
Neoliberal Optimism: Applying Market Techniques to Global Health.
Mei, Yuyang
2017-01-01
Global health and neoliberalism are becoming increasingly intertwined as organizations utilize markets and profit motives to solve the traditional problems of poverty and population health. I use field work conducted over 14 months in a global health technology company to explore how the promise of neoliberalism re-envisions humanitarian efforts. In this company's vaccine refrigerator project, staff members expect their investors and their market to allow them to achieve scale and develop accountability to their users in developing countries. However, the translation of neoliberal techniques to the global health sphere falls short of the ideal, as profits are meager and purchasing power remains with donor organizations. The continued optimism in market principles amidst such a non-ideal market reveals the tenacious ideological commitment to neoliberalism in these global health projects.
Solving Unconstrained Global Optimization Problems via Hybrid Swarm Intelligence Approaches
Directory of Open Access Journals (Sweden)
Jui-Yu Wu
2013-01-01
Full Text Available Stochastic global optimization (SGO algorithms such as the particle swarm optimization (PSO approach have become popular for solving unconstrained global optimization (UGO problems. The PSO approach, which belongs to the swarm intelligence domain, does not require gradient information, enabling it to overcome this limitation of traditional nonlinear programming methods. Unfortunately, PSO algorithm implementation and performance depend on several parameters, such as cognitive parameter, social parameter, and constriction coefficient. These parameters are tuned by using trial and error. To reduce the parametrization of a PSO method, this work presents two efficient hybrid SGO approaches, namely, a real-coded genetic algorithm-based PSO (RGA-PSO method and an artificial immune algorithm-based PSO (AIA-PSO method. The specific parameters of the internal PSO algorithm are optimized using the external RGA and AIA approaches, and then the internal PSO algorithm is applied to solve UGO problems. The performances of the proposed RGA-PSO and AIA-PSO algorithms are then evaluated using a set of benchmark UGO problems. Numerical results indicate that, besides their ability to converge to a global minimum for each test UGO problem, the proposed RGA-PSO and AIA-PSO algorithms outperform many hybrid SGO algorithms. Thus, the RGA-PSO and AIA-PSO approaches can be considered alternative SGO approaches for solving standard-dimensional UGO problems.
An ensemble-based method for constrained reservoir life-cycle optimization
Leeuwenburgh, O.; Egberts, P.J.P.; Chitu, A.G.
2015-01-01
We consider the problem of finding optimal long-term (life-cycle) recovery strategies for hydrocarbon reservoirs by use of simulation models. In such problems the presence of operating constraints, such as for example a maximum rate limit for a group of wells, may strongly influence the range of
Constrained Optimization Problems in Cost and Managerial Accounting--Spreadsheet Tools
Amlie, Thomas T.
2009-01-01
A common problem addressed in Managerial and Cost Accounting classes is that of selecting an optimal production mix given scarce resources. That is, if a firm produces a number of different products, and is faced with scarce resources (e.g., limitations on labor, materials, or machine time), what combination of products yields the greatest profit…
A policy iteration approach to online optimal control of continuous-time constrained-input systems.
Modares, Hamidreza; Naghibi Sistani, Mohammad-Bagher; Lewis, Frank L
2013-09-01
This paper is an effort towards developing an online learning algorithm to find the optimal control solution for continuous-time (CT) systems subject to input constraints. The proposed method is based on the policy iteration (PI) technique which has recently evolved as a major technique for solving optimal control problems. Although a number of online PI algorithms have been developed for CT systems, none of them take into account the input constraints caused by actuator saturation. In practice, however, ignoring these constraints leads to performance degradation or even system instability. In this paper, to deal with the input constraints, a suitable nonquadratic functional is employed to encode the constraints into the optimization formulation. Then, the proposed PI algorithm is implemented on an actor-critic structure to solve the Hamilton-Jacobi-Bellman (HJB) equation associated with this nonquadratic cost functional in an online fashion. That is, two coupled neural network (NN) approximators, namely an actor and a critic are tuned online and simultaneously for approximating the associated HJB solution and computing the optimal control policy. The critic is used to evaluate the cost associated with the current policy, while the actor is used to find an improved policy based on information provided by the critic. Convergence to a close approximation of the HJB solution as well as stability of the proposed feedback control law are shown. Simulation results of the proposed method on a nonlinear CT system illustrate the effectiveness of the proposed approach. Copyright © 2013 ISA. All rights reserved.
Directory of Open Access Journals (Sweden)
G. E. Bodeker
2013-02-01
Full Text Available High vertical resolution ozone measurements from eight different satellite-based instruments have been merged with data from the global ozonesonde network to calculate monthly mean ozone values in 5° latitude zones. These ''Tier 0'' ozone number densities and ozone mixing ratios are provided on 70 altitude levels (1 to 70 km and on 70 pressure levels spaced ~ 1 km apart (878.4 hPa to 0.046 hPa. The Tier 0 data are sparse and do not cover the entire globe or altitude range. To provide a gap-free database, a least squares regression model is fitted to the Tier 0 data and then evaluated globally. The regression model fit coefficients are expanded in Legendre polynomials to account for latitudinal structure, and in Fourier series to account for seasonality. Regression model fit coefficient patterns, which are two dimensional fields indexed by latitude and month of the year, from the N-th vertical level serve as an initial guess for the fit at the N + 1-th vertical level. The initial guess field for the first fit level (20 km/58.2 hPa was derived by applying the regression model to total column ozone fields. Perturbations away from the initial guess are captured through the Legendre and Fourier expansions. By applying a single fit at each level, and using the approach of allowing the regression fits to change only slightly from one level to the next, the regression is less sensitive to measurement anomalies at individual stations or to individual satellite-based instruments. Particular attention is paid to ensuring that the low ozone abundances in the polar regions are captured. By summing different combinations of contributions from different regression model basis functions, four different ''Tier 1'' databases have been compiled for different intended uses. This database is suitable for assessing ozone fields from chemistry-climate model simulations or for providing the ozone boundary conditions for global climate model simulations that do not
Natural gas fugitive emissions rates constrained by global atmospheric methane and ethane.
Schwietzke, Stefan; Griffin, W Michael; Matthews, H Scott; Bruhwiler, Lori M P
2014-07-15
The amount of methane emissions released by the natural gas (NG) industry is a critical and uncertain value for various industry and policy decisions, such as for determining the climate implications of using NG over coal. Previous studies have estimated fugitive emissions rates (FER)--the fraction of produced NG (mainly methane and ethane) escaped to the atmosphere--between 1 and 9%. Most of these studies rely on few and outdated measurements, and some may represent only temporal/regional NG industry snapshots. This study estimates NG industry representative FER using global atmospheric methane and ethane measurements over three decades, and literature ranges of (i) tracer gas atmospheric lifetimes, (ii) non-NG source estimates, and (iii) fossil fuel fugitive gas hydrocarbon compositions. The modeling suggests an upper bound global average FER of 5% during 2006-2011, and a most likely FER of 2-4% since 2000, trending downward. These results do not account for highly uncertain natural hydrocarbon seepage, which could lower the FER. Further emissions reductions by the NG industry may be needed to ensure climate benefits over coal during the next few decades.
Blind Channel Equalization with Colored Source Based on Constrained Optimization Methods
Directory of Open Access Journals (Sweden)
Dayong Zhou
2008-12-01
Full Text Available Tsatsanis and Xu have applied the constrained minimum output variance (CMOV principle to directly blind equalize a linear channelÃ¢Â€Â”a technique that has proven effective with white inputs. It is generally assumed in the literature that their CMOV method can also effectively equalize a linear channel with a colored source. In this paper, we prove that colored inputs will cause the equalizer to incorrectly converge due to inadequate constraints. We also introduce a new blind channel equalizer algorithm that is based on the CMOV principle, but with a different constraint that will correctly handle colored sources. Our proposed algorithm works for channels with either white or colored inputs and performs equivalently to the trained minimum mean-square error (MMSE equalizer under high SNR. Thus, our proposed algorithm may be regarded as an extension of the CMOV algorithm proposed by Tsatsanis and Xu. We also introduce several methods to improve the performance of our introduced algorithm in the low SNR condition. Simulation results show the superior performance of our proposed methods.
An Optimal Method for Developing Global Supply Chain Management System
Directory of Open Access Journals (Sweden)
Hao-Chun Lu
2013-01-01
Full Text Available Owing to the transparency in supply chains, enhancing competitiveness of industries becomes a vital factor. Therefore, many developing countries look for a possible method to save costs. In this point of view, this study deals with the complicated liberalization policies in the global supply chain management system and proposes a mathematical model via the flow-control constraints, which are utilized to cope with the bonded warehouses for obtaining maximal profits. Numerical experiments illustrate that the proposed model can be effectively solved to obtain the optimal profits in the global supply chain environment.
GLOBAL OPTIMIZATION METHODS FOR GRAVITATIONAL LENS SYSTEMS WITH REGULARIZED SOURCES
International Nuclear Information System (INIS)
Rogers, Adam; Fiege, Jason D.
2012-01-01
Several approaches exist to model gravitational lens systems. In this study, we apply global optimization methods to find the optimal set of lens parameters using a genetic algorithm. We treat the full optimization procedure as a two-step process: an analytical description of the source plane intensity distribution is used to find an initial approximation to the optimal lens parameters; the second stage of the optimization uses a pixelated source plane with the semilinear method to determine an optimal source. Regularization is handled by means of an iterative method and the generalized cross validation (GCV) and unbiased predictive risk estimator (UPRE) functions that are commonly used in standard image deconvolution problems. This approach simultaneously estimates the optimal regularization parameter and the number of degrees of freedom in the source. Using the GCV and UPRE functions, we are able to justify an estimation of the number of source degrees of freedom found in previous work. We test our approach by applying our code to a subset of the lens systems included in the SLACS survey.
Global Optimization of Minority Game by Smart Agents
Yan-Bo Xie; Bing-Hong Wang; Chin-Kun Hu; Tao Zhou
2004-01-01
We propose a new model of minority game with so-called smart agents such that the standard deviation and the total loss in this model reach the theoretical minimum values in the limit of long time. The smart agents use trail and error method to make a choice but bring global optimization to the system, which suggests that the economic systems may have the ability to self-organize into a highly optimized state by agents who are forced to make decisions based on inductive thinking for their lim...
Global crustal movement and tectonic plate boundary deformation constrained by the ITRF2008
Directory of Open Access Journals (Sweden)
Zhu Ze
2012-08-01
Full Text Available On the basis of the newly released International Terrestrial Reference Frame(ITRF2008 by the International Earth Rotation Service (IERS, a new global plate model ITRF2008 plate for the major plates is established. This ITRF2008-derived model is analyzed in comparison with NNR-NUVEL1A model, which is mainly based on geological and geophysical data. The Eurasia and Pacific plates display obvious differences in terms of the velocity fields derived from the two plate motion models. Plate acceleration is also introduced to characterize the differences of the two velocity fields which obtained from ITRF2008 -plate and NNR-NUVEL1A models for major individual plates. The results show that the Africa, South America and Eurasia plates are undergoing acceleration, while the North America and Australia plates are in the state of deceleration motion.
Lesmana, E.; Chaerani, D.; Khansa, H. N.
2018-03-01
Energy-Saving Generation Dispatch (ESGD) is a scheme made by Chinese Government in attempt to minimize CO2 emission produced by power plant. This scheme is made related to global warming which is primarily caused by too much CO2 in earth’s atmosphere, and while the need of electricity is something absolute, the power plants producing it are mostly thermal-power plant which produced many CO2. Many approach to fulfill this scheme has been made, one of them came through Minimum Cost Flow in which resulted in a Quadratically Constrained Quadratic Programming (QCQP) form. In this paper, ESGD problem with Minimum Cost Flow in QCQP form will be solved using Lagrange’s Multiplier Method
How does network design constrain optimal operation of intermittent water supply?
Lieb, Anna; Wilkening, Jon; Rycroft, Chris
2015-11-01
Urban water distribution systems do not always supply water continuously or reliably. As pipes fill and empty, pressure transients may contribute to degraded infrastructure and poor water quality. To help understand and manage this undesirable side effect of intermittent water supply--a phenomenon affecting hundreds of millions of people in cities around the world--we study the relative contributions of fixed versus dynamic properties of the network. Using a dynamical model of unsteady transition pipe flow, we study how different elements of network design, such as network geometry, pipe material, and pipe slope, contribute to undesirable pressure transients. Using an optimization framework, we then investigate to what extent network operation decisions such as supply timing and inflow rate may mitigate these effects. We characterize some aspects of network design that make them more or less amenable to operational optimization.
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Farouq, S.; Neytcheva, M.
2017-01-01
Roč. 310, January 2017 (2017), s. 5-18 ISSN 0377-0427 R&D Projects: GA MŠk ED1.1.00/02.0070 Institutional support: RVO:68145535 Keywords : optimal control * time-harmonic Stokes problem * preconditioning Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.357, year: 2016 http://www. science direct.com/ science /article/pii/S0377042716302631?via%3Dihub
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Farouq, S.; Neytcheva, M.
2017-01-01
Roč. 310, January 2017 (2017), s. 5-18 ISSN 0377-0427 R&D Projects: GA MŠk ED1.1.00/02.0070 Institutional support: RVO:68145535 Keywords : optimal control * time-harmonic Stokes problem * preconditioning Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 1.357, year: 2016 http://www.sciencedirect.com/science/article/pii/S0377042716302631?via%3Dihub
Agosta, John Mark
2013-01-01
This paper works through the optimization of a real world planning problem, with a combination of a generative planning tool and an influence diagram solver. The problem is taken from an existing application in the domain of oil spill emergency response. The planning agent manages constraints that order sets of feasible equipment employment actions. This is mapped at an intermediate level of abstraction onto an influence diagram. In addition, the planner can apply a surveillance operator that...
Directory of Open Access Journals (Sweden)
Yeison Díaz-Mateus
2017-07-01
Full Text Available Decision making in supply chains is influenced by demand variations, and hence sales, purchase orders and inventory levels are therefore concerned. This paper presents a non-linear optimization model for a two-echelon supply chain, for a unique product. In addition, the model includes the consumers’ maximum willingness to pay, taking socioeconomic differences into account. To do so, the constrained multinomial logit for discrete choices is used to estimate demand levels. Then, a metaheuristic approach based on particle swarm optimization is proposed to determine the optimal product sales price and inventory coordination variables. To validate the proposed model, a supply chain of a technological product was chosen and three scenarios are analyzed: discounts, demand segmentation and demand overestimation. Results are analyzed on the basis of profits, lotsizing and inventory turnover and market share. It can be concluded that the maximum willingness to pay must be taken into consideration, otherwise fictitious profits may mislead decision making, and although the market share would seem to improve, overall profits are not in fact necessarily better.
An Algorithm for Global Optimization Inspired by Collective Animal Behavior
Directory of Open Access Journals (Sweden)
Erik Cuevas
2012-01-01
Full Text Available A metaheuristic algorithm for global optimization called the collective animal behavior (CAB is introduced. Animal groups, such as schools of fish, flocks of birds, swarms of locusts, and herds of wildebeest, exhibit a variety of behaviors including swarming about a food source, milling around a central locations, or migrating over large distances in aligned groups. These collective behaviors are often advantageous to groups, allowing them to increase their harvesting efficiency, to follow better migration routes, to improve their aerodynamic, and to avoid predation. In the proposed algorithm, the searcher agents emulate a group of animals which interact with each other based on the biological laws of collective motion. The proposed method has been compared to other well-known optimization algorithms. The results show good performance of the proposed method when searching for a global optimum of several benchmark functions.
Cohen, J. B.
2015-12-01
The short lifetime and heterogeneous distribution of Black Carbon (BC) in the atmosphere leads to complex impacts on radiative forcing, climate, and health, and complicates analysis of its atmospheric processing and emissions. Two recent papers have estimated the global and regional emissions of BC using advanced statistical and computational methods. One used a Kalman Filter, including data from AERONET, NOAA, and other ground-based sources, to estimate global emissions of 17.8+/-5.6 Tg BC/year (with the increase attributable to East Asia, South Asia, Southeast Asia, and Eastern Europe - all regions which have had rapid urban, industrial, and economic expansion). The second additionally used remotely sensed measurements from MISR and a variance maximizing technique, uniquely quantifying fire and urban sources in Southeast Asia, as well as their large year-to-year variability over the past 12 years, leading to increases from 10% to 150%. These new emissions products, when run through our state-of-the art modelling system of chemistry, physics, transport, removal, radiation, and climate, match 140 ground stations and satellites better in both an absolute and a temporal sense. New work now further includes trace species measurements from OMI, which are used with the variance maximizing technique to constrain the types of emissions sources. Furthermore, land-use change and fire estimation products from MODIS are also included, which provide other constraints on the temporal and spatial nature of the variations of intermittent sources like fires or new permanent sources like expanded urbanization. This talk will introduce a new, top-down constrained, weekly varying BC emissions dataset, show that it produces a better fit with observations, and draw conclusions about the sources and impacts from urbanization one hand, and fires on another hand. Results specific to the Southeast and East Asia will demonstrate inter- and intra-annual variations, such as the function of
Optimal design of RTCs in digital circuit fault self-repair based on global signal optimization
Institute of Scientific and Technical Information of China (English)
Zhang Junbin; Cai Jinyan; Meng Yafeng
2016-01-01
Since digital circuits have been widely and thoroughly applied in various fields, electronic systems are increasingly more complicated and require greater reliability. Faults may occur in elec-tronic systems in complicated environments. If immediate field repairs are not made on the faults, elec-tronic systems will not run normally, and this will lead to serious losses. The traditional method for improving system reliability based on redundant fault-tolerant technique has been unable to meet the requirements. Therefore, on the basis of (evolvable hardware)-based and (reparation balance technology)-based electronic circuit fault self-repair strategy proposed in our preliminary work, the optimal design of rectification circuits (RTCs) in electronic circuit fault self-repair based on global sig-nal optimization is deeply researched in this paper. First of all, the basic theory of RTC optimal design based on global signal optimization is proposed. Secondly, relevant considerations and suitable ranges are analyzed. Then, the basic flow of RTC optimal design is researched. Eventually, a typical circuit is selected for simulation verification, and detailed simulated analysis is made on five circumstances that occur during RTC evolution. The simulation results prove that compared with the conventional design method based RTC, the global signal optimization design method based RTC is lower in hardware cost, faster in circuit evolution, higher in convergent precision, and higher in circuit evolution success rate. Therefore, the global signal optimization based RTC optimal design method applied in the elec-tronic circuit fault self-repair technology is proven to be feasible, effective, and advantageous.
Energy Technology Data Exchange (ETDEWEB)
Watkins, W.T.; Siebers, J.V. [University of Virginia, Charlottesville, VA (United States)
2016-06-15
Purpose: To introduce quasi-constrained Multi-Criteria Optimization (qcMCO) for unsupervised radiation therapy optimization which generates alternative patient-specific plans emphasizing dosimetric tradeoffs and conformance to clinical constraints for multiple delivery techniques. Methods: For N Organs At Risk (OARs) and M delivery techniques, qcMCO generates M(N+1) alternative treatment plans per patient. Objective weight variations for OARs and targets are used to generate alternative qcMCO plans. For 30 locally advanced lung cancer patients, qcMCO plans were generated for dosimetric tradeoffs to four OARs: each lung, heart, and esophagus (N=4) and 4 delivery techniques (simple 4-field arrangements, 9-field coplanar IMRT, 27-field non-coplanar IMRT, and non-coplanar Arc IMRT). Quasi-constrained objectives included target prescription isodose to 95% (PTV-D95), maximum PTV dose (PTV-Dmax)< 110% of prescription, and spinal cord Dmax<45 Gy. The algorithm’s ability to meet these constraints while simultaneously revealing dosimetric tradeoffs was investigated. Statistically significant dosimetric tradeoffs were defined such that the coefficient of determination between dosimetric indices which varied by at least 5 Gy between different plans was >0.8. Results: The qcMCO plans varied mean dose by >5 Gy to ipsilateral lung for 24/30 patients, contralateral lung for 29/30 patients, esophagus for 29/30 patients, and heart for 19/30 patients. In the 600 plans computed without human interaction, average PTV-D95=67.4±3.3 Gy, PTV-Dmax=79.2±5.3 Gy, and spinal cord Dmax was >45 Gy in 93 plans (>50 Gy in 2/600 plans). Statistically significant dosimetric tradeoffs were evident in 19/30 plans, including multiple tradeoffs of at least 5 Gy between multiple OARs in 7/30 cases. The most common statistically significant tradeoff was increasing PTV-Dmax to reduce OAR dose (15/30 patients). Conclusion: The qcMCO method can conform to quasi-constrained objectives while revealing
International Nuclear Information System (INIS)
Watkins, W.T.; Siebers, J.V.
2016-01-01
Purpose: To introduce quasi-constrained Multi-Criteria Optimization (qcMCO) for unsupervised radiation therapy optimization which generates alternative patient-specific plans emphasizing dosimetric tradeoffs and conformance to clinical constraints for multiple delivery techniques. Methods: For N Organs At Risk (OARs) and M delivery techniques, qcMCO generates M(N+1) alternative treatment plans per patient. Objective weight variations for OARs and targets are used to generate alternative qcMCO plans. For 30 locally advanced lung cancer patients, qcMCO plans were generated for dosimetric tradeoffs to four OARs: each lung, heart, and esophagus (N=4) and 4 delivery techniques (simple 4-field arrangements, 9-field coplanar IMRT, 27-field non-coplanar IMRT, and non-coplanar Arc IMRT). Quasi-constrained objectives included target prescription isodose to 95% (PTV-D95), maximum PTV dose (PTV-Dmax)< 110% of prescription, and spinal cord Dmax<45 Gy. The algorithm’s ability to meet these constraints while simultaneously revealing dosimetric tradeoffs was investigated. Statistically significant dosimetric tradeoffs were defined such that the coefficient of determination between dosimetric indices which varied by at least 5 Gy between different plans was >0.8. Results: The qcMCO plans varied mean dose by >5 Gy to ipsilateral lung for 24/30 patients, contralateral lung for 29/30 patients, esophagus for 29/30 patients, and heart for 19/30 patients. In the 600 plans computed without human interaction, average PTV-D95=67.4±3.3 Gy, PTV-Dmax=79.2±5.3 Gy, and spinal cord Dmax was >45 Gy in 93 plans (>50 Gy in 2/600 plans). Statistically significant dosimetric tradeoffs were evident in 19/30 plans, including multiple tradeoffs of at least 5 Gy between multiple OARs in 7/30 cases. The most common statistically significant tradeoff was increasing PTV-Dmax to reduce OAR dose (15/30 patients). Conclusion: The qcMCO method can conform to quasi-constrained objectives while revealing
Optimization of a constrained linear monochromator design for neutral atom beams
International Nuclear Information System (INIS)
Kaltenbacher, Thomas
2016-01-01
A focused ground state, neutral atom beam, exploiting its de Broglie wavelength by means of atom optics, is used for neutral atom microscopy imaging. Employing Fresnel zone plates as a lens for these beams is a well established microscopy technique. To date, even for favorable beam source conditions a minimal focus spot size of slightly below 1 μm was reached. This limitation is essentially given by the intrinsic spectral purity of the beam in combination with the chromatic aberration of the diffraction based zone plate. Therefore, it is important to enhance the monochromaticity of the beam, enabling a higher spatial resolution, preferably below 100 nm. We propose to increase the monochromaticity of a neutral atom beam by means of a so-called linear monochromator set-up – a Fresnel zone plate in combination with a pinhole aperture – in order to gain more than one order of magnitude in spatial resolution. This configuration is known in X-ray microscopy and has proven to be useful, but has not been applied to neutral atom beams. The main result of this work is optimal design parameters based on models for this linear monochromator set-up followed by a second zone plate for focusing. The optimization was performed for minimizing the focal spot size and maximizing the centre line intensity at the detector position for an atom beam simultaneously. The results presented in this work are for, but not limited to, a neutral helium atom beam. - Highlights: • The presented results are essential for optimal operation conditions of a neutral atom microscope set-up. • The key parameters for the experimental arrangement of a neutral microscopy set-up are identified and their interplay is quantified. • Insights in the multidimensional problem provide deep and crucial understanding for pushing beyond the apparent focus limitations. • This work points out the trade-offs for high intensity and high spatial resolution indicating several use cases.
Optimization of a constrained linear monochromator design for neutral atom beams
Energy Technology Data Exchange (ETDEWEB)
Kaltenbacher, Thomas
2016-04-15
A focused ground state, neutral atom beam, exploiting its de Broglie wavelength by means of atom optics, is used for neutral atom microscopy imaging. Employing Fresnel zone plates as a lens for these beams is a well established microscopy technique. To date, even for favorable beam source conditions a minimal focus spot size of slightly below 1 μm was reached. This limitation is essentially given by the intrinsic spectral purity of the beam in combination with the chromatic aberration of the diffraction based zone plate. Therefore, it is important to enhance the monochromaticity of the beam, enabling a higher spatial resolution, preferably below 100 nm. We propose to increase the monochromaticity of a neutral atom beam by means of a so-called linear monochromator set-up – a Fresnel zone plate in combination with a pinhole aperture – in order to gain more than one order of magnitude in spatial resolution. This configuration is known in X-ray microscopy and has proven to be useful, but has not been applied to neutral atom beams. The main result of this work is optimal design parameters based on models for this linear monochromator set-up followed by a second zone plate for focusing. The optimization was performed for minimizing the focal spot size and maximizing the centre line intensity at the detector position for an atom beam simultaneously. The results presented in this work are for, but not limited to, a neutral helium atom beam. - Highlights: • The presented results are essential for optimal operation conditions of a neutral atom microscope set-up. • The key parameters for the experimental arrangement of a neutral microscopy set-up are identified and their interplay is quantified. • Insights in the multidimensional problem provide deep and crucial understanding for pushing beyond the apparent focus limitations. • This work points out the trade-offs for high intensity and high spatial resolution indicating several use cases.
Energy Technology Data Exchange (ETDEWEB)
Tupek, Michael R. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
2016-06-30
In recent years there has been a proliferation of modeling techniques for forward predictions of crack propagation in brittle materials, including: phase-field/gradient damage models, peridynamics, cohesive-zone models, and G/XFEM enrichment techniques. However, progress on the corresponding inverse problems has been relatively lacking. Taking advantage of key features of existing modeling approaches, we propose a parabolic regularization of Barenblatt cohesive models which borrows extensively from previous phase-field and gradient damage formulations. An efficient explicit time integration strategy for this type of nonlocal fracture model is then proposed and justified. In addition, we present a C++ computational framework for computing in- put parameter sensitivities efficiently for explicit dynamic problems using the adjoint method. This capability allows for solving inverse problems involving crack propagation to answer interesting engineering questions such as: 1) what is the optimal design topology and material placement for a heterogeneous structure to maximize fracture resistance, 2) what loads must have been applied to a structure for it to have failed in an observed way, 3) what are the existing cracks in a structure given various experimental observations, etc. In this work, we focus on the first of these engineering questions and demonstrate a capability to automatically and efficiently compute optimal designs intended to minimize crack propagation in structures.
Proposal of Evolutionary Simplex Method for Global Optimization Problem
Shimizu, Yoshiaki
To make an agile decision in a rational manner, role of optimization engineering has been notified increasingly under diversified customer demand. With this point of view, in this paper, we have proposed a new evolutionary method serving as an optimization technique in the paradigm of optimization engineering. The developed method has prospects to solve globally various complicated problem appearing in real world applications. It is evolved from the conventional method known as Nelder and Mead’s Simplex method by virtue of idea borrowed from recent meta-heuristic method such as PSO. Mentioning an algorithm to handle linear inequality constraints effectively, we have validated effectiveness of the proposed method through comparison with other methods using several benchmark problems.
Globally optimal superconducting magnets part II: symmetric MSE coil arrangement.
Tieng, Quang M; Vegh, Viktor; Brereton, Ian M
2009-01-01
A globally optimal superconducting magnet coil design procedure based on the Minimum Stored Energy (MSE) current density map is outlined. The method has the ability to arrange coils in a manner that generates a strong and homogeneous axial magnetic field over a predefined region, and ensures the stray field external to the assembly and peak magnetic field at the wires are in acceptable ranges. The outlined strategy of allocating coils within a given domain suggests that coils should be placed around the perimeter of the domain with adjacent coils possessing alternating winding directions for optimum performance. The underlying current density maps from which the coils themselves are derived are unique, and optimized to possess minimal stored energy. Therefore, the method produces magnet designs with the lowest possible overall stored energy. Optimal coil layouts are provided for unshielded and shielded short bore symmetric superconducting magnets.
Global optimization for quantum dynamics of few-fermion systems
Li, Xikun; Pecak, Daniel; Sowiński, Tomasz; Sherson, Jacob; Nielsen, Anne E. B.
2018-03-01
Quantum state preparation is vital to quantum computation and quantum information processing tasks. In adiabatic state preparation, the target state is theoretically obtained with nearly perfect fidelity if the control parameter is tuned slowly enough. As this, however, leads to slow dynamics, it is often desirable to be able to carry out processes more rapidly. In this work, we employ two global optimization methods to estimate the quantum speed limit for few-fermion systems confined in a one-dimensional harmonic trap. Such systems can be produced experimentally in a well-controlled manner. We determine the optimized control fields and achieve a reduction in the ramping time of more than a factor of four compared to linear ramping. We also investigate how robust the fidelity is to small variations of the control fields away from the optimized shapes.
Lewis, Ian; Ottosen, Ann; Rubin, Jennifer; Blanc, Diana Chang; Zipursky, Simona; Wootton, Emily
2017-07-01
A total of 105 countries have introduced IPV as of September 2016 of which 85 have procured the vaccine through UNICEF. The Global Eradication and Endgame Strategic Plan 2013-2018 called for the rapid introduction of at least one dose of IPV into routine immunization schedules in 126 all OPV-using countries by the end of 2015. At the time of initiating the procurement process, demand was estimated based on global modeling rather than individual country indications. In its capacity as procurement agency for the Global Polio Eradication Initiative and Gavi, the Vaccine Alliance, UNICEF set out to secure access to IPV supply for around 100 countries. Based on offers received, sufficient supply was awarded to two manufacturers to meet projected routine requirements. However, due to technical issues scaling up vaccine production and an unforecasted demand for IPV use in campaigns to interrupt wild polio virus and to control type 2 vaccine derived polio virus outbreaks, IPV supplies are severely constrained. Activities to stretch supplies and to suppress demand have been ongoing since 2014, including delaying IPV introduction in countries where risks of type 2 reintroduction are lower, implementing the multi-dose vial policy, and encouraging the use of fractional dose delivered intradermally. Despite these efforts, there is still insufficient IPV supply to meet demand. The impact of the supply situation on IPV introduction timelines in countries are the focus of this article, and based on lessons learned with the IPV introductions, it is recommended for future health programs with accelerated scale up of programs, to take a cautious approach on supply commitments, putting in place clear allocation criteria in case of shortages or delays and establishing a communication strategy vis a vis beneficiaries. © The Author 2017. Published by Oxford University Press for the Infectious Diseases Society of America.
Ottosen, Ann; Rubin, Jennifer; Blanc, Diana Chang; Zipursky, Simona; Wootton, Emily
2017-01-01
Abstract A total of 105 countries have introduced IPV as of September 2016 of which 85 have procured the vaccine through UNICEF. The Global Eradication and Endgame Strategic Plan 2013-2018 called for the rapid introduction of at least one dose of IPV into routine immunization schedules in 126 all OPV-using countries by the end of 2015. At the time of initiating the procurement process, demand was estimated based on global modeling rather than individual country indications. In its capacity as procurement agency for the Global Polio Eradication Initiative and Gavi, the Vaccine Alliance, UNICEF set out to secure access to IPV supply for around 100 countries. Based on offers received, sufficient supply was awarded to two manufacturers to meet projected routine requirements. However, due to technical issues scaling up vaccine production and an unforecasted demand for IPV use in campaigns to interrupt wild polio virus and to control type 2 vaccine derived polio virus outbreaks, IPV supplies are severely constrained. Activities to stretch supplies and to suppress demand have been ongoing since 2014, including delaying IPV introduction in countries where risks of type 2 reintroduction are lower, implementing the multi-dose vial policy, and encouraging the use of fractional dose delivered intradermally. Despite these efforts, there is still insufficient IPV supply to meet demand. The impact of the supply situation on IPV introduction timelines in countries are the focus of this article, and based on lessons learned with the IPV introductions, it is recommended for future health programs with accelerated scale up of programs, to take a cautious approach on supply commitments, putting in place clear allocation criteria in case of shortages or delays and establishing a communication strategy vis a vis beneficiaries. PMID:28838159
A multi-fidelity analysis selection method using a constrained discrete optimization formulation
Stults, Ian C.
The purpose of this research is to develop a method for selecting the fidelity of contributing analyses in computer simulations. Model uncertainty is a significant component of result validity, yet it is neglected in most conceptual design studies. When it is considered, it is done so in only a limited fashion, and therefore brings the validity of selections made based on these results into question. Neglecting model uncertainty can potentially cause costly redesigns of concepts later in the design process or can even cause program cancellation. Rather than neglecting it, if one were to instead not only realize the model uncertainty in tools being used but also use this information to select the tools for a contributing analysis, studies could be conducted more efficiently and trust in results could be quantified. Methods for performing this are generally not rigorous or traceable, and in many cases the improvement and additional time spent performing enhanced calculations are washed out by less accurate calculations performed downstream. The intent of this research is to resolve this issue by providing a method which will minimize the amount of time spent conducting computer simulations while meeting accuracy and concept resolution requirements for results. In many conceptual design programs, only limited data is available for quantifying model uncertainty. Because of this data sparsity, traditional probabilistic means for quantifying uncertainty should be reconsidered. This research proposes to instead quantify model uncertainty using an evidence theory formulation (also referred to as Dempster-Shafer theory) in lieu of the traditional probabilistic approach. Specific weaknesses in using evidence theory for quantifying model uncertainty are identified and addressed for the purposes of the Fidelity Selection Problem. A series of experiments was conducted to address these weaknesses using n-dimensional optimization test functions. These experiments found that model
International Nuclear Information System (INIS)
Li Kaile; Ma Lijun
2005-01-01
We developed a source blocking optimization algorithm for Gamma Knife radiosurgery, which is based on tracking individual source contributions to arbitrarily shaped target and critical structure volumes. A scalar objective function and a direct search algorithm were used to produce near real-time calculation results. The algorithm allows the user to set and vary the total number of plugs for each shot to limit the total beam-on time. We implemented and tested the algorithm for several multiple-isocenter Gamma Knife cases. It was found that the use of limited number of plugs significantly lowered the integral dose to the critical structures such as an optical chiasm in pituitary adenoma cases. The main effect of the source blocking is the faster dose falloff in the junction area between the target and the critical structure. In summary, we demonstrated a useful source-plugging algorithm for improving complex multi-isocenter Gamma Knife treatment planning cases
DEFF Research Database (Denmark)
Tamas-Selicean, Domitian; Pop, Paul
2011-01-01
In this paper we are interested to implement mixed-criticality hard real-time applications on a given heterogeneous distributed architecture. Applications have different criticality levels, captured by their Safety-Integrity Level (SIL), and are scheduled using static-cyclic scheduling. Mixed......-criticality tasks can be integrated onto the same architecture only if there is enough spatial and temporal separation among them. We consider that the separation is provided by partitioning, such that applications run in separate partitions, and each partition is allocated several time slots on a processor. Tasks...... slots on each processor and (iv) the schedule tables, such that all the applications are schedulable and the development costs are minimized. We have proposed a Tabu Search-based approach to solve this optimization problem. The proposed algorithm has been evaluated using several synthetic and real...
Esfandiari, Kasra; Abdollahi, Farzaneh; Talebi, Heidar Ali
2017-09-01
In this paper, an identifier-critic structure is introduced to find an online near-optimal controller for continuous-time nonaffine nonlinear systems having saturated control signal. By employing two Neural Networks (NNs), the solution of Hamilton-Jacobi-Bellman (HJB) equation associated with the cost function is derived without requiring a priori knowledge about system dynamics. Weights of the identifier and critic NNs are tuned online and simultaneously such that unknown terms are approximated accurately and the control signal is kept between the saturation bounds. The convergence of NNs' weights, identification error, and system states is guaranteed using Lyapunov's direct method. Finally, simulation results are performed on two nonlinear systems to confirm the effectiveness of the proposed control strategy. Copyright © 2017 Elsevier Ltd. All rights reserved.
Optimization of a constrained linear monochromator design for neutral atom beams.
Kaltenbacher, Thomas
2016-04-01
A focused ground state, neutral atom beam, exploiting its de Broglie wavelength by means of atom optics, is used for neutral atom microscopy imaging. Employing Fresnel zone plates as a lens for these beams is a well established microscopy technique. To date, even for favorable beam source conditions a minimal focus spot size of slightly below 1μm was reached. This limitation is essentially given by the intrinsic spectral purity of the beam in combination with the chromatic aberration of the diffraction based zone plate. Therefore, it is important to enhance the monochromaticity of the beam, enabling a higher spatial resolution, preferably below 100nm. We propose to increase the monochromaticity of a neutral atom beam by means of a so-called linear monochromator set-up - a Fresnel zone plate in combination with a pinhole aperture - in order to gain more than one order of magnitude in spatial resolution. This configuration is known in X-ray microscopy and has proven to be useful, but has not been applied to neutral atom beams. The main result of this work is optimal design parameters based on models for this linear monochromator set-up followed by a second zone plate for focusing. The optimization was performed for minimizing the focal spot size and maximizing the centre line intensity at the detector position for an atom beam simultaneously. The results presented in this work are for, but not limited to, a neutral helium atom beam. Copyright © 2016 Elsevier B.V. All rights reserved.
Directory of Open Access Journals (Sweden)
Jiekun Song
2016-01-01
Full Text Available Harmonious development of 3Es (economy-energy-environment system is the key to realize regional sustainable development. The structure and components of 3Es system are analyzed. Based on the analysis of causality diagram, GDP and industrial structure are selected as the target parameters of economy subsystem, energy consumption intensity is selected as the target parameter of energy subsystem, and the emissions of COD, ammonia nitrogen, SO2, and NOX and CO2 emission intensity are selected as the target parameters of environment system. Fixed assets investment of three industries, total energy consumption, and investment in environmental pollution control are selected as the decision variables. By regarding the parameters of 3Es system optimization as fuzzy numbers, a fuzzy chance-constrained goal programming (FCCGP model is constructed, and a hybrid intelligent algorithm including fuzzy simulation and genetic algorithm is proposed for solving it. The results of empirical analysis on Shandong province of China show that the FCCGP model can reflect the inherent relationship and evolution law of 3Es system and provide the effective decision-making support for 3Es system optimization.
Xu, Y; Li, N
2014-09-01
Biological species have produced many simple but efficient rules in their complex and critical survival activities such as hunting and mating. A common feature observed in several biological motion strategies is that the predator only moves along paths in a carefully selected or iteratively refined subspace (or manifold), which might be able to explain why these motion strategies are effective. In this paper, a unified linear algebraic formulation representing such a predator-prey relationship is developed to simplify the construction and refinement process of the subspace (or manifold). Specifically, the following three motion strategies are studied and modified: motion camouflage, constant absolute target direction and local pursuit. The framework constructed based on this varying subspace concept could significantly reduce the computational cost in solving a class of nonlinear constrained optimal trajectory planning problems, particularly for the case with severe constraints. Two non-trivial examples, a ground robot and a hypersonic aircraft trajectory optimization problem, are used to show the capabilities of the algorithms in this new computational framework.
International Nuclear Information System (INIS)
Xu, Y; Li, N
2014-01-01
Biological species have produced many simple but efficient rules in their complex and critical survival activities such as hunting and mating. A common feature observed in several biological motion strategies is that the predator only moves along paths in a carefully selected or iteratively refined subspace (or manifold), which might be able to explain why these motion strategies are effective. In this paper, a unified linear algebraic formulation representing such a predator–prey relationship is developed to simplify the construction and refinement process of the subspace (or manifold). Specifically, the following three motion strategies are studied and modified: motion camouflage, constant absolute target direction and local pursuit. The framework constructed based on this varying subspace concept could significantly reduce the computational cost in solving a class of nonlinear constrained optimal trajectory planning problems, particularly for the case with severe constraints. Two non-trivial examples, a ground robot and a hypersonic aircraft trajectory optimization problem, are used to show the capabilities of the algorithms in this new computational framework. (paper)
Scheduling Multilevel Deadline-Constrained Scientific Workflows on Clouds Based on Cost Optimization
Directory of Open Access Journals (Sweden)
Maciej Malawski
2015-01-01
Full Text Available This paper presents a cost optimization model for scheduling scientific workflows on IaaS clouds such as Amazon EC2 or RackSpace. We assume multiple IaaS clouds with heterogeneous virtual machine instances, with limited number of instances per cloud and hourly billing. Input and output data are stored on a cloud object store such as Amazon S3. Applications are scientific workflows modeled as DAGs as in the Pegasus Workflow Management System. We assume that tasks in the workflows are grouped into levels of identical tasks. Our model is specified using mathematical programming languages (AMPL and CMPL and allows us to minimize the cost of workflow execution under deadline constraints. We present results obtained using our model and the benchmark workflows representing real scientific applications in a variety of domains. The data used for evaluation come from the synthetic workflows and from general purpose cloud benchmarks, as well as from the data measured in our own experiments with Montage, an astronomical application, executed on Amazon EC2 cloud. We indicate how this model can be used for scenarios that require resource planning for scientific workflows and their ensembles.
Global structural optimizations of surface systems with a genetic algorithm
International Nuclear Information System (INIS)
Chuang, Feng-Chuan
2005-01-01
Global structural optimizations with a genetic algorithm were performed for atomic cluster and surface systems including aluminum atomic clusters, Si magic clusters on the Si(111) 7 x 7 surface, silicon high-index surfaces, and Ag-induced Si(111) reconstructions. First, the global structural optimizations of neutral aluminum clusters Al n (n up to 23) were performed using a genetic algorithm coupled with a tight-binding potential. Second, a genetic algorithm in combination with tight-binding and first-principles calculations were performed to study the structures of magic clusters on the Si(111) 7 x 7 surface. Extensive calculations show that the magic cluster observed in scanning tunneling microscopy (STM) experiments consist of eight Si atoms. Simulated STM images of the Si magic cluster exhibit a ring-like feature similar to STM experiments. Third, a genetic algorithm coupled with a highly optimized empirical potential were used to determine the lowest energy structure of high-index semiconductor surfaces. The lowest energy structures of Si(105) and Si(114) were determined successfully. The results of Si(105) and Si(114) are reported within the framework of highly optimized empirical potential and first-principles calculations. Finally, a genetic algorithm coupled with Si and Ag tight-binding potentials were used to search for Ag-induced Si(111) reconstructions at various Ag and Si coverages. The optimized structural models of √3 x √3, 3 x 1, and 5 x 2 phases were reported using first-principles calculations. A novel model is found to have lower surface energy than the proposed double-honeycomb chained (DHC) model both for Au/Si(111) 5 x 2 and Ag/Si(111) 5 x 2 systems
Groenwold, A.A.; Wood, D.W.; Etman, L.F.P.; Tosserams, S.
2009-01-01
We implement and test a globally convergent sequential approximate optimization algorithm based on (convexified) diagonal quadratic approximations. The algorithm resides in the class of globally convergent optimization methods based on conservative convex separable approximations developed by
Effective Alternating Direction Optimization Methods for Sparsity-Constrained Blind Image Deblurring
Directory of Open Access Journals (Sweden)
Naixue Xiong
2017-01-01
Full Text Available Single-image blind deblurring for imaging sensors in the Internet of Things (IoT is a challenging ill-conditioned inverse problem, which requires regularization techniques to stabilize the image restoration process. The purpose is to recover the underlying blur kernel and latent sharp image from only one blurred image. Under many degraded imaging conditions, the blur kernel could be considered not only spatially sparse, but also piecewise smooth with the support of a continuous curve. By taking advantage of the hybrid sparse properties of the blur kernel, a hybrid regularization method is proposed in this paper to robustly and accurately estimate the blur kernel. The effectiveness of the proposed blur kernel estimation method is enhanced by incorporating both the L 1 -norm of kernel intensity and the squared L 2 -norm of the intensity derivative. Once the accurate estimation of the blur kernel is obtained, the original blind deblurring can be simplified to the direct deconvolution of blurred images. To guarantee robust non-blind deconvolution, a variational image restoration model is presented based on the L 1 -norm data-fidelity term and the total generalized variation (TGV regularizer of second-order. All non-smooth optimization problems related to blur kernel estimation and non-blind deconvolution are effectively handled by using the alternating direction method of multipliers (ADMM-based numerical methods. Comprehensive experiments on both synthetic and realistic datasets have been implemented to compare the proposed method with several state-of-the-art methods. The experimental comparisons have illustrated the satisfactory imaging performance of the proposed method in terms of quantitative and qualitative evaluations.
A Novel Hybrid Firefly Algorithm for Global Optimization.
Directory of Open Access Journals (Sweden)
Lina Zhang
Full Text Available Global optimization is challenging to solve due to its nonlinearity and multimodality. Traditional algorithms such as the gradient-based methods often struggle to deal with such problems and one of the current trends is to use metaheuristic algorithms. In this paper, a novel hybrid population-based global optimization algorithm, called hybrid firefly algorithm (HFA, is proposed by combining the advantages of both the firefly algorithm (FA and differential evolution (DE. FA and DE are executed in parallel to promote information sharing among the population and thus enhance searching efficiency. In order to evaluate the performance and efficiency of the proposed algorithm, a diverse set of selected benchmark functions are employed and these functions fall into two groups: unimodal and multimodal. The experimental results show better performance of the proposed algorithm compared to the original version of the firefly algorithm (FA, differential evolution (DE and particle swarm optimization (PSO in the sense of avoiding local minima and increasing the convergence rate.
A Global Network Alignment Method Using Discrete Particle Swarm Optimization.
Huang, Jiaxiang; Gong, Maoguo; Ma, Lijia
2016-10-19
Molecular interactions data increase exponentially with the advance of biotechnology. This makes it possible and necessary to comparatively analyse the different data at a network level. Global network alignment is an important network comparison approach to identify conserved subnetworks and get insight into evolutionary relationship across species. Network alignment which is analogous to subgraph isomorphism is known to be an NP-hard problem. In this paper, we introduce a novel heuristic Particle-Swarm-Optimization based Network Aligner (PSONA), which optimizes a weighted global alignment model considering both protein sequence similarity and interaction conservations. The particle statuses and status updating rules are redefined in a discrete form by using permutation. A seed-and-extend strategy is employed to guide the searching for the superior alignment. The proposed initialization method "seeds" matches with high sequence similarity into the alignment, which guarantees the functional coherence of the mapping nodes. A greedy local search method is designed as the "extension" procedure to iteratively optimize the edge conservations. PSONA is compared with several state-of-art methods on ten network pairs combined by five species. The experimental results demonstrate that the proposed aligner can map the proteins with high functional coherence and can be used as a booster to effectively refine the well-studied aligners.
Paasche, H.; Tronicke, J.
2012-04-01
In many near surface geophysical applications multiple tomographic data sets are routinely acquired to explore subsurface structures and parameters. Linking the model generation process of multi-method geophysical data sets can significantly reduce ambiguities in geophysical data analysis and model interpretation. Most geophysical inversion approaches rely on local search optimization methods used to find an optimal model in the vicinity of a user-given starting model. The final solution may critically depend on the initial model. Alternatively, global optimization (GO) methods have been used to invert geophysical data. They explore the solution space in more detail and determine the optimal model independently from the starting model. Additionally, they can be used to find sets of optimal models allowing a further analysis of model parameter uncertainties. Here we employ particle swarm optimization (PSO) to realize the global optimization of tomographic data. PSO is an emergent methods based on swarm intelligence characterized by fast and robust convergence towards optimal solutions. The fundamental principle of PSO is inspired by nature, since the algorithm mimics the behavior of a flock of birds searching food in a search space. In PSO, a number of particles cruise a multi-dimensional solution space striving to find optimal model solutions explaining the acquired data. The particles communicate their positions and success and direct their movement according to the position of the currently most successful particle of the swarm. The success of a particle, i.e. the quality of the currently found model by a particle, must be uniquely quantifiable to identify the swarm leader. When jointly inverting disparate data sets, the optimization solution has to satisfy multiple optimization objectives, at least one for each data set. Unique determination of the most successful particle currently leading the swarm is not possible. Instead, only statements about the Pareto
The energy transition in a climate-constrained world: Regional vs. global optimization
Brede, M.; de Vries, B.
2013-01-01
In this paper we present a stylized economy-energy-climate model and discuss the role of the atmosphere, fossil fuels, and a stock of accumulated knowledge about renewable energy technologies in collaboratively and competitively managed worlds. The model highlights that assumptions about the 'degree
National Research Council Canada - National Science Library
Dennis, John E; El-Alem, Mahmoud; Maciel, Maria C
1995-01-01
.... The tangential component then must satisfy a fraction of Cauchy decrease condition of a quadratic model of the Lagrasigian function in the translated tangent space of the constraints determined by the quasi-normal component.
National Research Council Canada - National Science Library
Dennis, John E; El-Alem, Mahmoud; Maciel, Maria C
1995-01-01
.... The normal Component need not be computed accurately. The theory requires a quasi-normal component to satisfy a fraction of Cauchy decrease condition on the quadratic model of the linearized constraints...
Model-data fusion across ecosystems: from multisite optimizations to global simulations
Kuppel, S.; Peylin, P.; Maignan, F.; Chevallier, F.; Kiely, G.; Montagnani, L.; Cescatti, A.
2014-11-01
This study uses a variational data assimilation framework to simultaneously constrain a global ecosystem model with eddy covariance measurements of daily net ecosystem exchange (NEE) and latent heat (LE) fluxes from a large number of sites grouped in seven plant functional types (PFTs). It is an attempt to bridge the gap between the numerous site-specific parameter optimization works found in the literature and the generic parameterization used by most land surface models within each PFT. The present multisite approach allows deriving PFT-generic sets of optimized parameters enhancing the agreement between measured and simulated fluxes at most of the sites considered, with performances often comparable to those of the corresponding site-specific optimizations. Besides reducing the PFT-averaged model-data root-mean-square difference (RMSD) and the associated daily output uncertainty, the optimization improves the simulated CO2 balance at tropical and temperate forests sites. The major site-level NEE adjustments at the seasonal scale are reduced amplitude in C3 grasslands and boreal forests, increased seasonality in temperate evergreen forests, and better model-data phasing in temperate deciduous broadleaf forests. Conversely, the poorer performances in tropical evergreen broadleaf forests points to deficiencies regarding the modelling of phenology and soil water stress for this PFT. An evaluation with data-oriented estimates of photosynthesis (GPP - gross primary productivity) and ecosystem respiration (Reco) rates indicates distinctively improved simulations of both gross fluxes. The multisite parameter sets are then tested against CO2 concentrations measured at 53 locations around the globe, showing significant adjustments of the modelled seasonality of atmospheric CO2 concentration, whose relevance seems PFT-dependent, along with an improved interannual variability. Lastly, a global-scale evaluation with remote sensing NDVI (normalized difference vegetation index
Globally optimal superconducting magnets part I: minimum stored energy (MSE) current density map.
Tieng, Quang M; Vegh, Viktor; Brereton, Ian M
2009-01-01
An optimal current density map is crucial in magnet design to provide the initial values within search spaces in an optimization process for determining the final coil arrangement of the magnet. A strategy for obtaining globally optimal current density maps for the purpose of designing magnets with coaxial cylindrical coils in which the stored energy is minimized within a constrained domain is outlined. The current density maps obtained utilising the proposed method suggests that peak current densities occur around the perimeter of the magnet domain, where the adjacent peaks have alternating current directions for the most compact designs. As the dimensions of the domain are increased, the current density maps yield traditional magnet designs of positive current alone. These unique current density maps are obtained by minimizing the stored magnetic energy cost function and therefore suggest magnet coil designs of minimal system energy. Current density maps are provided for a number of different domain arrangements to illustrate the flexibility of the method and the quality of the achievable designs.
Global optimization of minority game by intelligent agents
Xie, Yan-Bo; Wang, Bing-Hong; Hu, Chin-Kun; Zhou, Tao
2005-10-01
We propose a new model of minority game with intelligent agents who use trail and error method to make a choice such that the standard deviation σ2 and the total loss in this model reach the theoretical minimum values in the long time limit and the global optimization of the system is reached. This suggests that the economic systems can self-organize into a highly optimized state by agents who make decisions based on inductive thinking, limited knowledge, and capabilities. When other kinds of agents are also present, the simulation results and analytic calculations show that the intelligent agent can gain profits from producers and are much more competent than the noise traders and conventional agents in original minority games proposed by Challet and Zhang.
An Adaptive Unified Differential Evolution Algorithm for Global Optimization
Energy Technology Data Exchange (ETDEWEB)
Qiang, Ji; Mitchell, Chad
2014-11-03
In this paper, we propose a new adaptive unified differential evolution algorithm for single-objective global optimization. Instead of the multiple mutation strate- gies proposed in conventional differential evolution algorithms, this algorithm employs a single equation unifying multiple strategies into one expression. It has the virtue of mathematical simplicity and also provides users the flexibility for broader exploration of the space of mutation operators. By making all control parameters in the proposed algorithm self-adaptively evolve during the process of optimization, it frees the application users from the burden of choosing appro- priate control parameters and also improves the performance of the algorithm. In numerical tests using thirteen basic unimodal and multimodal functions, the proposed adaptive unified algorithm shows promising performance in compari- son to several conventional differential evolution algorithms.
Global optimization in the adaptive assay of subterranean uranium nodules
International Nuclear Information System (INIS)
Vulkan, U.; Ben-Haim, Y.
1989-01-01
An adaptive assay is one in which the design of the assay system is modified during operation in response to measurements obtained on-line. The present work has two aims: to design an adaptive system for borehole assay of isolated subterranean uranium nodules, and to investigate globality of optimal design in adaptive assay. It is shown experimentally that reasonably accurate estimates of uranium mass are obtained for a wide range of nodule shapes, on the basis of an adaptive assay system based on a simple geomorphological model. Furthermore, two concepts are identified which underlie the optimal design of the assay system. The adaptive assay approach shows promise for successful measurement of spatially random material in many geophysical applications. (author)
A concept for global optimization of topology design problems
DEFF Research Database (Denmark)
Stolpe, Mathias; Achtziger, Wolfgang; Kawamoto, Atsushi
2006-01-01
We present a concept for solving topology design problems to proven global optimality. We propose that the problems are modeled using the approach of simultaneous analysis and design with discrete design variables and solved with convergent branch and bound type methods. This concept is illustrated...... on two applications. The first application is the design of stiff truss structures where the bar areas are chosen from a finite set of available areas. The second considered application is simultaneous topology and geometry design of planar articulated mechanisms. For each application we outline...
A Unified Differential Evolution Algorithm for Global Optimization
Energy Technology Data Exchange (ETDEWEB)
Qiang, Ji; Mitchell, Chad
2014-06-24
Abstract?In this paper, we propose a new unified differential evolution (uDE) algorithm for single objective global optimization. Instead of selecting among multiple mutation strategies as in the conventional differential evolution algorithm, this algorithm employs a single equation as the mutation strategy. It has the virtue of mathematical simplicity and also provides users the flexbility for broader exploration of different mutation strategies. Numerical tests using twelve basic unimodal and multimodal functions show promising performance of the proposed algorithm in comparison to convential differential evolution algorithms.
Simulated Annealing-Based Krill Herd Algorithm for Global Optimization
Directory of Open Access Journals (Sweden)
Gai-Ge Wang
2013-01-01
Full Text Available Recently, Gandomi and Alavi proposed a novel swarm intelligent method, called krill herd (KH, for global optimization. To enhance the performance of the KH method, in this paper, a new improved meta-heuristic simulated annealing-based krill herd (SKH method is proposed for optimization tasks. A new krill selecting (KS operator is used to refine krill behavior when updating krill’s position so as to enhance its reliability and robustness dealing with optimization problems. The introduced KS operator involves greedy strategy and accepting few not-so-good solutions with a low probability originally used in simulated annealing (SA. In addition, a kind of elitism scheme is used to save the best individuals in the population in the process of the krill updating. The merits of these improvements are verified by fourteen standard benchmarking functions and experimental results show that, in most cases, the performance of this improved meta-heuristic SKH method is superior to, or at least highly competitive with, the standard KH and other optimization methods.
Palacios, S.G.
2015-01-01
In health facilities in resource-constrained settings, a lack of access to sustainable and reliable electricity can result on a sub-optimal delivery of healthcare services, as they do not have lighting for medical procedures and power to run essential equipment and devices to treat their patients.
Regis, Rommel G.
2014-02-01
This article develops two new algorithms for constrained expensive black-box optimization that use radial basis function surrogates for the objective and constraint functions. These algorithms are called COBRA and Extended ConstrLMSRBF and, unlike previous surrogate-based approaches, they can be used for high-dimensional problems where all initial points are infeasible. They both follow a two-phase approach where the first phase finds a feasible point while the second phase improves this feasible point. COBRA and Extended ConstrLMSRBF are compared with alternative methods on 20 test problems and on the MOPTA08 benchmark automotive problem (D.R. Jones, Presented at MOPTA 2008), which has 124 decision variables and 68 black-box inequality constraints. The alternatives include a sequential penalty derivative-free algorithm, a direct search method with kriging surrogates, and two multistart methods. Numerical results show that COBRA algorithms are competitive with Extended ConstrLMSRBF and they generally outperform the alternatives on the MOPTA08 problem and most of the test problems.
Directory of Open Access Journals (Sweden)
Kyungsung An
2017-05-01
Full Text Available This research aims to improve the operational efficiency and security of electric power systems at high renewable penetration by exploiting the envisioned controllability or flexibility of electric vehicles (EVs; EVs interact with the grid through grid-to-vehicle (G2V and vehicle-to-grid (V2G services to ensure reliable and cost-effective grid operation. This research provides a computational framework for this decision-making process. Charging and discharging strategies of EV aggregators are incorporated into a security-constrained optimal power flow (SCOPF problem such that overall energy cost is minimized and operation within acceptable reliability criteria is ensured. Particularly, this SCOPF problem has been formulated for Jeju Island in South Korea, in order to lower carbon emissions toward a zero-carbon island by, for example, integrating large-scale renewable energy and EVs. On top of conventional constraints on the generators and line flows, a unique constraint on the system inertia constant, interpreted as the minimum synchronous generation, is considered to ensure grid security at high renewable penetration. The available energy constraint of the participating EV associated with the state-of-charge (SOC of the battery and market price-responsive behavior of the EV aggregators are also explored. Case studies for the Jeju electric power system in 2030 under various operational scenarios demonstrate the effectiveness of the proposed method and improved operational flexibility via controllable EVs.
Klein, E.; Masson, F.; Duputel, Z.; Yavasoglu, H.; Agram, P. S.
2016-12-01
Over the last two decades, the densification of GPS networks and the development of new radar satellites offered an unprecedented opportunity to study crustal deformation due to faulting. Yet, submarine strike slip fault segments remain a major issue, especially when the landscape appears unfavorable to the use of SAR measurements. It is the case of the North Anatolian fault segments located in the Main Marmara Sea, that remain unbroken ever since the Mw7.4 earthquake of Izmit in 1999, which ended a eastward migrating seismic sequence of Mw > 7 earthquakes. Located directly offshore Istanbul, evaluation of seismic hazard appears capital. But a strong controversy remains over whether these segments are accumulating strain and are likely to experience a major earthquake, or are creeping, resulting both from the simplicity of current geodetic models and the scarcity of geodetic data. We indeed show that 2D infinite fault models cannot account for the complexity of the Marmara fault segments. But current geodetic data in the western region of Istanbul are also insufficient to invert for the coupling using a 3D geometry of the fault. Therefore, we implement a global optimization procedure aiming at identifying the most favorable distribution of GPS stations to explore the strain accumulation. We present here the results of this procedure that allows to determine both the optimal number and location of the new stations. We show that a denser terrestrial survey network can indeed locally improve the resolution on the shallower part of the fault, even more efficiently with permanent stations. But data closer from the fault, only possible by submarine measurements, remain necessary to properly constrain the fault behavior and its potential along strike coupling variations.
Zhang, Yong-Feng; Chiang, Hsiao-Dong
2017-09-01
A novel three-stage methodology, termed the "consensus-based particle swarm optimization (PSO)-assisted Trust-Tech methodology," to find global optimal solutions for nonlinear optimization problems is presented. It is composed of Trust-Tech methods, consensus-based PSO, and local optimization methods that are integrated to compute a set of high-quality local optimal solutions that can contain the global optimal solution. The proposed methodology compares very favorably with several recently developed PSO algorithms based on a set of small-dimension benchmark optimization problems and 20 large-dimension test functions from the CEC 2010 competition. The analytical basis for the proposed methodology is also provided. Experimental results demonstrate that the proposed methodology can rapidly obtain high-quality optimal solutions that can contain the global optimal solution. The scalability of the proposed methodology is promising.
Identification of metabolic system parameters using global optimization methods
Directory of Open Access Journals (Sweden)
Gatzke Edward P
2006-01-01
Full Text Available Abstract Background The problem of estimating the parameters of dynamic models of complex biological systems from time series data is becoming increasingly important. Methods and results Particular consideration is given to metabolic systems that are formulated as Generalized Mass Action (GMA models. The estimation problem is posed as a global optimization task, for which novel techniques can be applied to determine the best set of parameter values given the measured responses of the biological system. The challenge is that this task is nonconvex. Nonetheless, deterministic optimization techniques can be used to find a global solution that best reconciles the model parameters and measurements. Specifically, the paper employs branch-and-bound principles to identify the best set of model parameters from observed time course data and illustrates this method with an existing model of the fermentation pathway in Saccharomyces cerevisiae. This is a relatively simple yet representative system with five dependent states and a total of 19 unknown parameters of which the values are to be determined. Conclusion The efficacy of the branch-and-reduce algorithm is illustrated by the S. cerevisiae example. The method described in this paper is likely to be widely applicable in the dynamic modeling of metabolic networks.
GMG: A Guaranteed, Efficient Global Optimization Algorithm for Remote Sensing.
Energy Technology Data Exchange (ETDEWEB)
D' Helon, CD
2004-08-18
The monocular passive ranging (MPR) problem in remote sensing consists of identifying the precise range of an airborne target (missile, plane, etc.) from its observed radiance. This inverse problem may be set as a global optimization problem (GOP) whereby the difference between the observed and model predicted radiances is minimized over the possible ranges and atmospheric conditions. Using additional information about the error function between the predicted and observed radiances of the target, we developed GMG, a new algorithm to find the Global Minimum with a Guarantee. The new algorithm transforms the original continuous GOP into a discrete search problem, thereby guaranteeing to find the position of the global minimum in a reasonably short time. The algorithm is first applied to the golf course problem, which serves as a litmus test for its performance in the presence of both complete and degraded additional information. GMG is further assessed on a set of standard benchmark functions and then applied to various realizations of the MPR problem.
Spatiotemporal radiotherapy planning using a global optimization approach
Adibi, Ali; Salari, Ehsan
2018-02-01
This paper aims at quantifying the extent of potential therapeutic gain, measured using biologically effective dose (BED), that can be achieved by altering the radiation dose distribution over treatment sessions in fractionated radiotherapy. To that end, a spatiotemporally integrated planning approach is developed, where the spatial and temporal dose modulations are optimized simultaneously. The concept of equivalent uniform BED (EUBED) is used to quantify and compare the clinical quality of spatiotemporally heterogeneous dose distributions in target and critical structures. This gives rise to a large-scale non-convex treatment-plan optimization problem, which is solved using global optimization techniques. The proposed spatiotemporal planning approach is tested on two stylized cancer cases resembling two different tumor sites and sensitivity analysis is performed for radio-biological and EUBED parameters. Numerical results validate that spatiotemporal plans are capable of delivering a larger BED to the target volume without increasing the BED in critical structures compared to conventional time-invariant plans. In particular, this additional gain is attributed to the irradiation of different regions of the target volume at different treatment sessions. Additionally, the trade-off between the potential therapeutic gain and the number of distinct dose distributions is quantified, which suggests a diminishing marginal gain as the number of dose distributions increases.
Optimal correction and design parameter search by modern methods of rigorous global optimization
International Nuclear Information System (INIS)
Makino, K.; Berz, M.
2011-01-01
Frequently the design of schemes for correction of aberrations or the determination of possible operating ranges for beamlines and cells in synchrotrons exhibit multitudes of possibilities for their correction, usually appearing in disconnected regions of parameter space which cannot be directly qualified by analytical means. In such cases, frequently an abundance of optimization runs are carried out, each of which determines a local minimum depending on the specific chosen initial conditions. Practical solutions are then obtained through an often extended interplay of experienced manual adjustment of certain suitable parameters and local searches by varying other parameters. However, in a formal sense this problem can be viewed as a global optimization problem, i.e. the determination of all solutions within a certain range of parameters that lead to a specific optimum. For example, it may be of interest to find all possible settings of multiple quadrupoles that can achieve imaging; or to find ahead of time all possible settings that achieve a particular tune; or to find all possible manners to adjust nonlinear parameters to achieve correction of high order aberrations. These tasks can easily be phrased in terms of such an optimization problem; but while mathematically this formulation is often straightforward, it has been common belief that it is of limited practical value since the resulting optimization problem cannot usually be solved. However, recent significant advances in modern methods of rigorous global optimization make these methods feasible for optics design for the first time. The key ideas of the method lie in an interplay of rigorous local underestimators of the objective functions, and by using the underestimators to rigorously iteratively eliminate regions that lie above already known upper bounds of the minima, in what is commonly known as a branch-and-bound approach. Recent enhancements of the Differential Algebraic methods used in particle
DEFF Research Database (Denmark)
Achtziger, Wolfgang; Stolpe, Mathias
2007-01-01
this problem is well-studied for continuous bar areas, we consider in this study the case of discrete areas. This problem is of major practical relevance if the truss must be built from pre-produced bars with given areas. As a special case, we consider the design problem for a single available bar area, i.......e., a 0/1 problem. In contrast to the heuristic methods considered in many other approaches, our goal is to compute guaranteed globally optimal structures. This is done by a branch-and-bound method for which convergence can be proven. In this branch-and-bound framework, lower bounds of the optimal......-integer problems. The main intention of this paper is to provide optimal solutions for single and multiple load benchmark examples, which can be used for testing and validating other methods or heuristics for the treatment of this discrete topology design problem....
Global optimization applied to GPS positioning by ambiguity functions
International Nuclear Information System (INIS)
Baselga, Sergio
2010-01-01
Differential GPS positioning with carrier-phase observables is commonly done in a process that involves determination of the unknown integer ambiguity values. An alternative approach, named the ambiguity function method, was already proposed in the early days of GPS positioning. By making use of a trigonometric function ambiguity unknowns are eliminated from the functional model before the estimation process. This approach has significant advantages, such as ease of use and insensitivity to cycle slips, but requires such high accuracy in the initial approximate coordinates that its use has been practically dismissed from consideration. In this paper a novel strategy is proposed so that the need for highly accurate initial coordinates disappears: the application of a global optimization method to the ambiguity functions model. The use of this strategy enables the ambiguity function method to compete with the present prevailing approach of ambiguity resolution
Global optimization numerical strategies for rate-independent processes
Czech Academy of Sciences Publication Activity Database
Benešová, Barbora
2011-01-01
Roč. 50, č. 2 (2011), s. 197-220 ISSN 0925-5001 R&D Projects: GA ČR GAP201/10/0357 Grant - others:GA MŠk(CZ) LC06052 Program:LC Institutional research plan: CEZ:AV0Z20760514 Keywords : rate-independent processes * numerical global optimization * energy estimates based algorithm Subject RIV: BA - General Mathematics Impact factor: 1.196, year: 2011 http://math.hnue.edu.vn/portal/rss.viewpage.php?id=0000037780&ap=L3BvcnRhbC9ncmFiYmVyLnBocD9jYXRpZD0xMDEyJnBhZ2U9Mg==
Adjusting process count on demand for petascale global optimization
Sosonkina, Masha; Watson, Layne T.; Radcliffe, Nicholas R.; Haftka, Rafael T.; Trosset, Michael W.
2013-01-01
There are many challenges that need to be met before efficient and reliable computation at the petascale is possible. Many scientific and engineering codes running at the petascale are likely to be memory intensive, which makes thrashing a serious problem for many petascale applications. One way to overcome this challenge is to use a dynamic number of processes, so that the total amount of memory available for the computation can be increased on demand. This paper describes modifications made to the massively parallel global optimization code pVTdirect in order to allow for a dynamic number of processes. In particular, the modified version of the code monitors memory use and spawns new processes if the amount of available memory is determined to be insufficient. The primary design challenges are discussed, and performance results are presented and analyzed.
WFH: closing the global gap--achieving optimal care.
Skinner, Mark W
2012-07-01
For 50 years, the World Federation of Hemophilia (WFH) has been working globally to close the gap in care and to achieve Treatment for All patients, men and women, with haemophilia and other inherited bleeding disorders, regardless of where they might live. The WFH estimates that more than one in 1000 men and women has a bleeding disorder equating to 6,900,000 worldwide. To close the gap in care between developed and developing nations a continued focus on the successful strategies deployed heretofore will be required. However, in response to the rapid advances in treatment and emerging therapeutic advances on the horizon it will also require fresh approaches and renewed strategic thinking. It is difficult to predict what each therapeutic advance on the horizon will mean for the future, but there is no doubt that we are in a golden age of research and development, which has the prospect of revolutionizing treatment once again. An improved understanding of "optimal" treatment is fundamental to the continued evolution of global care. The challenges of answering government and payer demands for evidence-based medicine, and cost justification for the introduction and enhancement of treatment, are ever-present and growing. To sustain and improve care it is critical to build the body of outcome data for individual patients, within haemophilia treatment centers (HTCs), nationally, regionally and globally. Emerging therapeutic advances (longer half-life therapies and gene transfer) should not be justified or brought to market based only on the notion that they will be economically more affordable, although that may be the case, but rather more importantly that they will be therapeutically more advantageous. Improvements in treatment adherence, reductions in bleeding frequency (including microhemorrhages), better management of trough levels, and improved health outcomes (including quality of life) should be the foremost considerations. As part of a new WFH strategic plan
A practical globalization of one-shot optimization for optimal design of tokamak divertors
Energy Technology Data Exchange (ETDEWEB)
Blommaert, Maarten, E-mail: maarten.blommaert@kuleuven.be [Institute of Energy and Climate Research (IEK-4), FZ Jülich GmbH, D-52425 Jülich (Germany); Dekeyser, Wouter; Baelmans, Martine [KU Leuven, Department of Mechanical Engineering, 3001 Leuven (Belgium); Gauger, Nicolas R. [TU Kaiserslautern, Chair for Scientific Computing, 67663 Kaiserslautern (Germany); Reiter, Detlev [Institute of Energy and Climate Research (IEK-4), FZ Jülich GmbH, D-52425 Jülich (Germany)
2017-01-01
In past studies, nested optimization methods were successfully applied to design of the magnetic divertor configuration in nuclear fusion reactors. In this paper, so-called one-shot optimization methods are pursued. Due to convergence issues, a globalization strategy for the one-shot solver is sought. Whereas Griewank introduced a globalization strategy using a doubly augmented Lagrangian function that includes primal and adjoint residuals, its practical usability is limited by the necessity of second order derivatives and expensive line search iterations. In this paper, a practical alternative is offered that avoids these drawbacks by using a regular augmented Lagrangian merit function that penalizes only state residuals. Additionally, robust rank-two Hessian estimation is achieved by adaptation of Powell's damped BFGS update rule. The application of the novel one-shot approach to magnetic divertor design is considered in detail. For this purpose, the approach is adapted to be complementary with practical in parts adjoint sensitivities. Using the globalization strategy, stable convergence of the one-shot approach is achieved.
ABCluster: the artificial bee colony algorithm for cluster global optimization.
Zhang, Jun; Dolg, Michael
2015-10-07
Global optimization of cluster geometries is of fundamental importance in chemistry and an interesting problem in applied mathematics. In this work, we introduce a relatively new swarm intelligence algorithm, i.e. the artificial bee colony (ABC) algorithm proposed in 2005, to this field. It is inspired by the foraging behavior of a bee colony, and only three parameters are needed to control it. We applied it to several potential functions of quite different nature, i.e., the Coulomb-Born-Mayer, Lennard-Jones, Morse, Z and Gupta potentials. The benchmarks reveal that for long-ranged potentials the ABC algorithm is very efficient in locating the global minimum, while for short-ranged ones it is sometimes trapped into a local minimum funnel on a potential energy surface of large clusters. We have released an efficient, user-friendly, and free program "ABCluster" to realize the ABC algorithm. It is a black-box program for non-experts as well as experts and might become a useful tool for chemists to study clusters.
Arasomwan, Martins Akugbe; Adewumi, Aderemi Oluyinka
2013-01-01
Linear decreasing inertia weight (LDIW) strategy was introduced to improve on the performance of the original particle swarm optimization (PSO). However, linear decreasing inertia weight PSO (LDIW-PSO) algorithm is known to have the shortcoming of premature convergence in solving complex (multipeak) optimization problems due to lack of enough momentum for particles to do exploitation as the algorithm approaches its terminal point. Researchers have tried to address this shortcoming by modifying LDIW-PSO or proposing new PSO variants. Some of these variants have been claimed to outperform LDIW-PSO. The major goal of this paper is to experimentally establish the fact that LDIW-PSO is very much efficient if its parameters are properly set. First, an experiment was conducted to acquire a percentage value of the search space limits to compute the particle velocity limits in LDIW-PSO based on commonly used benchmark global optimization problems. Second, using the experimentally obtained values, five well-known benchmark optimization problems were used to show the outstanding performance of LDIW-PSO over some of its competitors which have in the past claimed superiority over it. Two other recent PSO variants with different inertia weight strategies were also compared with LDIW-PSO with the latter outperforming both in the simulation experiments conducted. PMID:24324383
Bacour, C.; Maignan, F.; Porcar-Castell, A.; MacBean, N.; Goulas, Y.; Flexas, J.; Guanter, L.; Joiner, J.; Peylin, P.
2016-12-01
A new era for improving our knowledge of the terrestrial carbon cycle at the global scale has begun with recent studies on the relationships between remotely sensed Sun Induce Fluorescence (SIF) and plant photosynthetic activity (GPP), and the availability of such satellite-derived products now "routinely" produced from GOSAT, GOME-2, or OCO-2 observations. Assimilating SIF data into terrestrial ecosystem models (TEMs) represents a novel opportunity to reduce the uncertainty of their prediction with respect to carbon-climate feedbacks, in particular the uncertainties resulting from inaccurate parameter values. A prerequisite is a correct representation in TEMs of the several drivers of plant fluorescence from the leaf to the canopy scale, and in particular the competing processes of photochemistry and non photochemical quenching (NPQ).In this study, we present the first results of a global scale assimilation of GOME-2 SIF products within a new version of the ORCHIDEE land surface model including a physical module of plant fluorescence. At the leaf level, the regulation of fluorescence yield is simulated both by the photosynthesis module of ORCHIDEE to calculate the photochemical yield and by a parametric model to estimate NPQ. The latter has been calibrated on leaf fluorescence measurements performed for boreal coniferous and Mediterranean vegetation species. A parametric representation of the SCOPE radiative transfer model is used to model the plant fluorescence fluxes for PSI and PSII and the scaling up to the canopy level. The ORCHIDEE-FluOR model is firstly evaluated with respect to in situ measurements of plant fluorescence flux and photochemical yield for scots pine and wheat. The potentials of SIF data to constrain the modelled GPP are evaluated by assimilating one year of GOME-2-SIF products within ORCHIDEE-FluOR. We investigate in particular the changes in the spatial patterns of GPP following the optimization of the photosynthesis and phenology parameters
Directory of Open Access Journals (Sweden)
Wei Li
2015-01-01
Full Text Available We propose a new optimization algorithm inspired by the formation and change of the cloud in nature, referred to as Cloud Particles Differential Evolution (CPDE algorithm. The cloud is assumed to have three states in the proposed algorithm. Gaseous state represents the global exploration. Liquid state represents the intermediate process from the global exploration to the local exploitation. Solid state represents the local exploitation. The best solution found so far acts as a nucleus. In gaseous state, the nucleus leads the population to explore by condensation operation. In liquid state, cloud particles carry out macrolocal exploitation by liquefaction operation. A new mutation strategy called cloud differential mutation is introduced in order to solve a problem that the misleading effect of a nucleus may cause the premature convergence. In solid state, cloud particles carry out microlocal exploitation by solidification operation. The effectiveness of the algorithm is validated upon different benchmark problems. The results have been compared with eight well-known optimization algorithms. The statistical analysis on performance evaluation of the different algorithms on 10 benchmark functions and CEC2013 problems indicates that CPDE attains good performance.
Optimizing Orbit-Instrument Configuration for Global Precipitation Mission (GPM) Satellite Fleet
Smith, Eric A.; Adams, James; Baptista, Pedro; Haddad, Ziad; Iguchi, Toshio; Im, Eastwood; Kummerow, Christian; Einaudi, Franco (Technical Monitor)
2001-01-01
architecture. Following this introduction, we focus specifically on the last topic, that being an analysis which leads to an optimal flight architecture dictated in part by science requirements but constrained by allowable orbital mechanics, instrument scan patterns, and antenna aperture properties. Because the optimal architecture involves an interplay between orbit mechanics and instrument specifications, it is important to recognize that in attempting to serve various scientific themes, the final optimal architecture will represent a compromise concerning dynamic range, spatial resolution, sampling interval, pointing, beam coincidence, and measurement uncertainty. Moreover, cost becomes a major factor in seeking the optimal architecture through the pathways of antenna and instrument scan designs, as well as propulsion requirements associated with the orbit heights of various "constellation" members. Although the results presented at the IGARSS-2001 meeting will likely not be the fully refined flight architecture specifications, they are expected to be nearly complete.
Automatic Construction and Global Optimization of a Multisentiment Lexicon
Directory of Open Access Journals (Sweden)
Xiaoping Yang
2016-01-01
Full Text Available Manual annotation of sentiment lexicons costs too much labor and time, and it is also difficult to get accurate quantification of emotional intensity. Besides, the excessive emphasis on one specific field has greatly limited the applicability of domain sentiment lexicons (Wang et al., 2010. This paper implements statistical training for large-scale Chinese corpus through neural network language model and proposes an automatic method of constructing a multidimensional sentiment lexicon based on constraints of coordinate offset. In order to distinguish the sentiment polarities of those words which may express either positive or negative meanings in different contexts, we further present a sentiment disambiguation algorithm to increase the flexibility of our lexicon. Lastly, we present a global optimization framework that provides a unified way to combine several human-annotated resources for learning our 10-dimensional sentiment lexicon SentiRuc. Experiments show the superior performance of SentiRuc lexicon in category labeling test, intensity labeling test, and sentiment classification tasks. It is worth mentioning that, in intensity label test, SentiRuc outperforms the second place by 21 percent.
A DE-Based Scatter Search for Global Optimization Problems
Directory of Open Access Journals (Sweden)
Kun Li
2015-01-01
Full Text Available This paper proposes a hybrid scatter search (SS algorithm for continuous global optimization problems by incorporating the evolution mechanism of differential evolution (DE into the reference set updated procedure of SS to act as the new solution generation method. This hybrid algorithm is called a DE-based SS (SSDE algorithm. Since different kinds of mutation operators of DE have been proposed in the literature and they have shown different search abilities for different kinds of problems, four traditional mutation operators are adopted in the hybrid SSDE algorithm. To adaptively select the mutation operator that is most appropriate to the current problem, an adaptive mechanism for the candidate mutation operators is developed. In addition, to enhance the exploration ability of SSDE, a reinitialization method is adopted to create a new population and subsequently construct a new reference set whenever the search process of SSDE is trapped in local optimum. Computational experiments on benchmark problems show that the proposed SSDE is competitive or superior to some state-of-the-art algorithms in the literature.
DEFF Research Database (Denmark)
Bech, Michael Møller; Nørgård, Christian; Roemer, Daniel Beck
2016-01-01
This paper illustrates how the relatively simple constrained multi-objective optimization algorithm Generalized Differential Evolution 3 (GDE3), can assist with the practical sizing of mechatronic components used in e.g. digital displacement fluid power machinery. The studied bi- and tri-objectiv......This paper illustrates how the relatively simple constrained multi-objective optimization algorithm Generalized Differential Evolution 3 (GDE3), can assist with the practical sizing of mechatronic components used in e.g. digital displacement fluid power machinery. The studied bi- and tri...... different optimization control parameter settings and it is concluded that GDE3 is a reliable optimization tool that can assist mechatronic engineers in the design and decision making process....
International Nuclear Information System (INIS)
Walrand, Stephan; Jamar, François; Pauwels, Stanislas
2009-01-01
Ill-posed linear systems occur in many different fields. A class of regularization methods, called constrained optimization, aims to determine the extremum of a penalty function whilst constraining an objective function to a likely value. We propose here a novel heuristic way to screen the local extrema satisfying the discrepancy principle. A modified version of the Landweber algorithm is used for the iteration process. After finding a local extremum, a bound is performed to the 'farthest' estimate in the data space still satisfying the discrepancy principle. Afterwards, the modified Landweber algorithm is again applied to find a new local extremum. This bound-iteration process is repeated until a satisfying solution is reached. For evaluation in nuclear medicine tomography, a novel penalty function that preserves the edge steps in the reconstructed solution was evaluated on Monte Carlo simulations and using real SPECT acquisitions as well. Surprisingly, the first bound always provided a significantly better solution in a wide range of statistics
Global Optimization Employing Gaussian Process-Based Bayesian Surrogates
Directory of Open Access Journals (Sweden)
Roland Preuss
2018-03-01
Full Text Available The simulation of complex physics models may lead to enormous computer running times. Since the simulations are expensive it is necessary to exploit the computational budget in the best possible manner. If for a few input parameter settings an output data set has been acquired, one could be interested in taking these data as a basis for finding an extremum and possibly an input parameter set for further computer simulations to determine it—a task which belongs to the realm of global optimization. Within the Bayesian framework we utilize Gaussian processes for the creation of a surrogate model function adjusted self-consistently via hyperparameters to represent the data. Although the probability distribution of the hyperparameters may be widely spread over phase space, we make the assumption that only the use of their expectation values is sufficient. While this shortcut facilitates a quickly accessible surrogate, it is somewhat justified by the fact that we are not interested in a full representation of the model by the surrogate but to reveal its maximum. To accomplish this the surrogate is fed to a utility function whose extremum determines the new parameter set for the next data point to obtain. Moreover, we propose to alternate between two utility functions—expected improvement and maximum variance—in order to avoid the drawbacks of each. Subsequent data points are drawn from the model function until the procedure either remains in the points found or the surrogate model does not change with the iteration. The procedure is applied to mock data in one and two dimensions in order to demonstrate proof of principle of the proposed approach.
Global stability-based design optimization of truss structures using ...
Indian Academy of Sciences (India)
Furthermore, a pure pareto-ranking based multi-objective optimization model is employed for the design optimization of the truss structure with multiple objectives. The computational performance of the optimization model is increased by implementing an island model into its evolutionary search mechanism. The proposed ...
Directory of Open Access Journals (Sweden)
Feng Zou
2016-01-01
Full Text Available An improved teaching-learning-based optimization with combining of the social character of PSO (TLBO-PSO, which is considering the teacher’s behavior influence on the students and the mean grade of the class, is proposed in the paper to find the global solutions of function optimization problems. In this method, the teacher phase of TLBO is modified; the new position of the individual is determined by the old position, the mean position, and the best position of current generation. The method overcomes disadvantage that the evolution of the original TLBO might stop when the mean position of students equals the position of the teacher. To decrease the computation cost of the algorithm, the process of removing the duplicate individual in original TLBO is not adopted in the improved algorithm. Moreover, the probability of local convergence of the improved method is decreased by the mutation operator. The effectiveness of the proposed method is tested on some benchmark functions, and the results are competitive with respect to some other methods.
Kerschke, Pascal
2017-01-01
Choosing the best-performing optimizer(s) out of a portfolio of optimization algorithms is usually a difficult and complex task. It gets even worse, if the underlying functions are unknown, i.e., so-called Black-Box problems, and function evaluations are considered to be expensive. In the case of continuous single-objective optimization problems, Exploratory Landscape Analysis (ELA) - a sophisticated and effective approach for characterizing the landscapes of such problems by means of numeric...
International Nuclear Information System (INIS)
Casas, E.; Troeltzsch, F.
1999-01-01
In this paper we are concerned with some optimal control problems governed by semilinear elliptic equations. The case of a boundary control is studied. We consider pointwise constraints on the control and a finite number of equality and inequality constraints on the state. The goal is to derive first- and second-order optimality conditions satisfied by locally optimal solutions of the problem
A Simple But Effective Canonical Dual Theory Unified Algorithm for Global Optimization
Zhang, Jiapu
2011-01-01
Numerical global optimization methods are often very time consuming and could not be applied for high-dimensional nonconvex/nonsmooth optimization problems. Due to the nonconvexity/nonsmoothness, directly solving the primal problems sometimes is very difficult. This paper presents a very simple but very effective canonical duality theory (CDT) unified global optimization algorithm. This algorithm has convergence is proved in this paper. More important, for this CDT-unified algorithm, numerous...
Effective Energy Methods for Global Optimization for Biopolymer Structure Prediction
National Research Council Canada - National Science Library
Shalloway, David
1998-01-01
.... Its main strength is that it uncovers and exploits the intrinsic "hidden structures" of biopolymer energy landscapes to efficiently perform global minimization using a hierarchical search procedure...
Jones, Keith
2010-01-01
The Regularized Fast Hartley Transform provides the reader with the tools necessary to both understand the proposed new formulation and to implement simple design variations that offer clear implementational advantages, both practical and theoretical, over more conventional complex-data solutions to the problem. The highly-parallel formulation described is shown to lead to scalable and device-independent solutions to the latency-constrained version of the problem which are able to optimize the use of the available silicon resources, and thus to maximize the achievable computational density, th
Directory of Open Access Journals (Sweden)
Daniel W. O’Neill
2016-01-01
Full Text Available Given the global disease burden and resource disparity that exists in the world and the globalization of Christianity, Christians are in a critical position to effect radical change in individuals, communities and systems for human flourishing. This paper describes the theological basis of seven key elements that the Church can contribute to sustainable development, global health equity, and universal access as an expanding movement: defining health, speaking truth, providing care, making peace, cooperating, setting priorities, and mobilizing resources for maximum stewardship in low resource settings.
International Nuclear Information System (INIS)
An, Y; Liang, J; Liu, W
2015-01-01
Purpose: We propose to apply a probabilistic framework, namely chanceconstrained optimization, in the intensity-modulated proton therapy (IMPT) planning subject to range and patient setup uncertainties. The purpose is to hedge against the influence of uncertainties and improve robustness of treatment plans. Methods: IMPT plans were generated for a typical prostate patient. Nine dose distributions are computed — the nominal one and one each for ±5mm setup uncertainties along three cardinal axes and for ±3.5% range uncertainty. These nine dose distributions are supplied to the solver CPLEX as chance constraints to explicitly control plan robustness under these representative uncertainty scenarios with certain probability. This probability is determined by the tolerance level. We make the chance-constrained model tractable by converting it to a mixed integer optimization problem. The quality of plans derived from this method is evaluated using dose-volume histogram (DVH) indices such as tumor dose homogeneity (D5% – D95%) and coverage (D95%) and normal tissue sparing like V70 of rectum, V65, and V40 of bladder. We also compare the results from this novel method with the conventional PTV-based method to further demonstrate its effectiveness Results: Our model can yield clinically acceptable plans within 50 seconds. The chance-constrained optimization produces IMPT plans with comparable target coverage, better target dose homogeneity, and better normal tissue sparing compared to the PTV-based optimization [D95% CTV: 67.9 vs 68.7 (Gy), D5% – D95% CTV: 11.9 vs 18 (Gy), V70 rectum: 0.0 % vs 0.33%, V65 bladder: 2.17% vs 9.33%, V40 bladder: 8.83% vs 21.83%]. It also simultaneously makes the plan more robust [Width of DVH band at D50%: 2.0 vs 10.0 (Gy)]. The tolerance level may be varied to control the tradeoff between plan robustness and quality. Conclusion: The chance-constrained optimization generates superior IMPT plan compared to the PTV-based optimization with
An Evaluation of the Sniffer Global Optimization Algorithm Using Standard Test Functions
Butler, Roger A. R.; Slaminka, Edward E.
1992-03-01
The performance of Sniffer—a new global optimization algorithm—is compared with that of Simulated Annealing. Using the number of function evaluations as a measure of efficiency, the new algorithm is shown to be significantly better at finding the global minimum of seven standard test functions. Several of the test functions used have many local minima and very steep walls surrounding the global minimum. Such functions are intended to thwart global minimization algorithms.
Directory of Open Access Journals (Sweden)
Baofeng Cai
2017-08-01
Full Text Available The Interconnected River System Network Project (IRSNP is a significant water supply engineering project, which is capable of effectively utilizing flood resources to generate ecological value, by connecting 198 lakes and ponds in western Jilin, northeast China. In this article, an optimization research approach has been proposed to maximize the incremental value of IRSNP ecosystem services. A double-sided chance-constrained integer linear program (DCCILP method has been proposed to support the optimization, which can deal with uncertainties presented as integers or random parameters that appear on both sides of the decision variable at the same time. The optimal scheme indicates that after rational optimization, the total incremental value of ecosystem services from the interconnected river system network project increased 22.25%, providing an increase in benefits of 3.26 × 109 ¥ compared to the original scheme. Most of the functional area is swamp wetland, which provides the greatest ecological benefits. Adjustment services increased obviously, implying that the optimization scheme prioritizes ecological benefits rather than supply and production services.
Global optimization for overall HVAC systems - Part I problem formulation and analysis
International Nuclear Information System (INIS)
Lu Lu; Cai Wenjian; Chai, Y.S.; Xie Lihua
2005-01-01
This paper presents the global optimization technologies for overall heating, ventilating and air conditioning (HVAC) systems. The objective function of global optimization and constraints are formulated based on mathematical models of the major components. All these models are associated with power consumption components and heat exchangers for transferring cooling load. The characteristics of all the major components are briefly introduced by models, and the interactions between them are analyzed and discussed to show the complications of the problem. According to the characteristics of the operating components, the complicated original optimization problem for overall HVAC systems is transformed and simplified into a compact form ready for optimization
International Nuclear Information System (INIS)
Maraver, Daniel; Royo, Javier; Lemort, Vincent; Quoilin, Sylvain
2014-01-01
Highlights: • ORC optimization for different target applications. • Model developed to allow computation in subcritical and transcritical operation. • Regenerative and non-regenerative cycles evaluated through second law efficiency. • Common working fluids: R134a, R245fa, Solkatherm, n-Pentane, MDM, Toluene. • Thermodynamic and technological approaches lead to optimal design guidelines. - Abstract: The present work is focused on the thermodynamic optimization of organic Rankine cycles (ORCs) for power generation and CHP from different average heat source profiles (waste heat recovery, thermal oil for cogeneration and geothermal). The general goal is to provide optimization guidelines for a wide range of operating conditions, for subcritical and transcritical, regenerative and non-regenerative cycles. A parameter assessment of the main equipment in the cycle (expander, heat exchangers and feed pump) was also carried out. An optimization model of the ORC (available as an electronic annex) is proposed to predict the best cycle performance (subcritical or transcritical), in terms of its exergy efficiency, with different working fluids. The working fluids considered are those most commonly used in commercial ORC units (R134a, R245fa, Solkatherm, n-Pentane, Octamethyltrisiloxane and Toluene). The optimal working fluid and operating conditions from a purely thermodynamic approach are limited by the technological constraints of the expander, the heat exchangers and the feed pump. Hence, a complementary assessment of both approaches is more adequate to obtain some preliminary design guidelines for ORC units
A branch and bound algorithm for the global optimization of Hessian Lipschitz continuous functions
Fowkes, Jaroslav M.; Gould, Nicholas I. M.; Farmer, Chris L.
2012-01-01
We present a branch and bound algorithm for the global optimization of a twice differentiable nonconvex objective function with a Lipschitz continuous Hessian over a compact, convex set. The algorithm is based on applying cubic regularisation
International Nuclear Information System (INIS)
Martorell, S.; Serradell, V.; Munoz, A.; Sanchez, A.
1997-01-01
Background, objective, scope, detailed working plan and follow-up and final product of the project ''Global optimization of maintenance and surveillance testing based on reliability and probabilistic safety assessment'' are described
Xu, Gang; Li, Ming; Mourrain, Bernard; Rabczuk, Timon; Xu, Jinlan; Bordas, Stéphane P. A.
2018-01-01
In this paper, we propose a general framework for constructing IGA-suitable planar B-spline parameterizations from given complex CAD boundaries consisting of a set of B-spline curves. Instead of forming the computational domain by a simple boundary, planar domains with high genus and more complex boundary curves are considered. Firstly, some pre-processing operations including B\\'ezier extraction and subdivision are performed on each boundary curve in order to generate a high-quality planar parameterization; then a robust planar domain partition framework is proposed to construct high-quality patch-meshing results with few singularities from the discrete boundary formed by connecting the end points of the resulting boundary segments. After the topology information generation of quadrilateral decomposition, the optimal placement of interior B\\'ezier curves corresponding to the interior edges of the quadrangulation is constructed by a global optimization method to achieve a patch-partition with high quality. Finally, after the imposition of C1=G1-continuity constraints on the interface of neighboring B\\'ezier patches with respect to each quad in the quadrangulation, the high-quality B\\'ezier patch parameterization is obtained by a C1-constrained local optimization method to achieve uniform and orthogonal iso-parametric structures while keeping the continuity conditions between patches. The efficiency and robustness of the proposed method are demonstrated by several examples which are compared to results obtained by the skeleton-based parameterization approach.
DEFF Research Database (Denmark)
Sørensen, Søren N.; Stolpe, Mathias
2015-01-01
rate. The capabilities of the method and the effect of active versus inactive manufacturing constraints are demonstrated on several numerical examples of limited size, involving at most 320 binary variables. Most examples are solved to guaranteed global optimality and may constitute benchmark examples...... but is, however, convex in the original mixed binary nested form. Convexity is the foremost important property of optimization problems, and the proposed method can guarantee the global or near-global optimal solution; unlike most topology optimization methods. The material selection is limited...... for popular topology optimization methods and heuristics based on solving sequences of non-convex problems. The results will among others demonstrate that the difficulty of the posed problem is highly dependent upon the composition of the constitutive properties of the material candidates....
Globally Optimal Segmentation of Permanent-Magnet Systems
DEFF Research Database (Denmark)
Insinga, Andrea Roberto; Bjørk, Rasmus; Smith, Anders
2016-01-01
Permanent-magnet systems are widely used for generation of magnetic fields with specific properties. The reciprocity theorem, an energy-equivalence principle in magnetostatics, can be employed to calculate the optimal remanent flux density of the permanent-magnet system, given any objective...... remains unsolved. We show that the problem of optimal segmentation of a two-dimensional permanent-magnet assembly with respect to a linear objective functional can be reduced to the problem of piecewise linear approximation of a plane curve by perimeter maximization. Once the problem has been cast...
The Tunneling Method for Global Optimization in Multidimensional Scaling.
Groenen, Patrick J. F.; Heiser, Willem J.
1996-01-01
A tunneling method for global minimization in multidimensional scaling is introduced and adjusted for multidimensional scaling with general Minkowski distances. The method alternates a local search step with a tunneling step in which a different configuration is sought with the same STRESS implementation. (SLD)
Global Optimization of a Periodic System using a Genetic Algorithm
Stucke, David; Crespi, Vincent
2001-03-01
We use a novel application of a genetic algorithm global optimizatin technique to find the lowest energy structures for periodic systems. We apply this technique to colloidal crystals for several different stoichiometries of binary and trinary colloidal crystals. This application of a genetic algorithm is decribed and results of likely candidate structures are presented.
Global Launcher Trajectory Optimization for Lunar Base Settlement
Pagano, A.; Mooij, E.
2010-01-01
The problem of a mission to the Moon to set a permanent outpost can be tackled by dividing the journey into three phases: the Earth ascent, the Earth-Moon transfer and the lunar landing. In this paper we present an optimization analysis of Earth ascent trajectories of existing launch vehicles
Vertical bifacial solar farms: Physics, design, and global optimization
Khan, M. Ryyan; Hanna, Amir; Sun, Xingshu; Alam, Muhammad A.
2017-01-01
10–20% more energy than a traditional monofacial farm for a practical row-spacing of 2 m (corresponding to 1.2 m high panels). With the prospect of additional 5–20% energy gain from reduced soiling and tilt optimization, bifacial solar farm do offer a
Global stability-based design optimization of truss structures using ...
Indian Academy of Sciences (India)
The quality of current pareto front obtained in the end of a whole genetic search is assessed according to its closeness to the ...... better optimal designation with a lower displacement value of 0.3075 in. satisfying the service- .... Internal force. R.
DEFF Research Database (Denmark)
Jing, Lishuai; Pedersen, Troels; Fleury, Bernard Henri
2013-01-01
for which we propose a genetic algorithm that computes close-to-optimal solutions. Simulation results demonstrate that the proposed algorithm can efficiently find pilot signals that outperform the state-of-the-art pilot signals in both single-path and multipath propagation scenarios. In addition, we...
Saborido, Rubén; Ruiz, Ana B; Luque, Mariano
2017-01-01
In this article, we propose a new evolutionary algorithm for multiobjective optimization called Global WASF-GA ( global weighting achievement scalarizing function genetic algorithm), which falls within the aggregation-based evolutionary algorithms. The main purpose of Global WASF-GA is to approximate the whole Pareto optimal front. Its fitness function is defined by an achievement scalarizing function (ASF) based on the Tchebychev distance, in which two reference points are considered (both utopian and nadir objective vectors) and the weight vector used is taken from a set of weight vectors whose inverses are well-distributed. At each iteration, all individuals are classified into different fronts. Each front is formed by the solutions with the lowest values of the ASF for the different weight vectors in the set, using the utopian vector and the nadir vector as reference points simultaneously. Varying the weight vector in the ASF while considering the utopian and the nadir vectors at the same time enables the algorithm to obtain a final set of nondominated solutions that approximate the whole Pareto optimal front. We compared Global WASF-GA to MOEA/D (different versions) and NSGA-II in two-, three-, and five-objective problems. The computational results obtained permit us to conclude that Global WASF-GA gets better performance, regarding the hypervolume metric and the epsilon indicator, than the other two algorithms in many cases, especially in three- and five-objective problems.
Sorzano, Carlos Oscars S; Pérez-De-La-Cruz Moreno, Maria Angeles; Burguet-Castell, Jordi; Montejo, Consuelo; Ros, Antonio Aguilar
2015-06-01
Pharmacokinetics (PK) applications can be seen as a special case of nonlinear, causal systems with memory. There are cases in which prior knowledge exists about the distribution of the system parameters in a population. However, for a specific patient in a clinical setting, we need to determine her system parameters so that the therapy can be personalized. This system identification is performed many times by measuring drug concentrations in plasma. The objective of this work is to provide an irregular sampling strategy that minimizes the uncertainty about the system parameters with a fixed amount of samples (cost constrained). We use Monte Carlo simulations to estimate the average Fisher's information matrix associated to the PK problem, and then estimate the sampling points that minimize the maximum uncertainty associated to system parameters (a minimax criterion). The minimization is performed employing a genetic algorithm. We show that such a sampling scheme can be designed in a way that is adapted to a particular patient and that it can accommodate any dosing regimen as well as it allows flexible therapeutic strategies. © 2015 Wiley Periodicals, Inc. and the American Pharmacists Association.
Avoiding spurious submovement decompositions: a globally optimal algorithm
International Nuclear Information System (INIS)
Rohrer, Brandon Robinson; Hogan, Neville
2003-01-01
Evidence for the existence of discrete submovements underlying continuous human movement has motivated many attempts to extract them. Although they produce visually convincing results, all of the methodologies that have been employed are prone to produce spurious decompositions. Examples of potential failures are given. A branch-and-bound algorithm for submovement extraction, capable of global nonlinear minimization (and hence capable of avoiding spurious decompositions), is developed and demonstrated.
A Global Optimization Algorithm for Sum of Linear Ratios Problem
Yuelin Gao; Siqiao Jin
2013-01-01
We equivalently transform the sum of linear ratios programming problem into bilinear programming problem, then by using the linear characteristics of convex envelope and concave envelope of double variables product function, linear relaxation programming of the bilinear programming problem is given, which can determine the lower bound of the optimal value of original problem. Therefore, a branch and bound algorithm for solving sum of linear ratios programming problem is put forward, and the c...
Global Optimization for Transport Network Expansion and Signal Setting
Liu, Haoxiang; Wang, David Z. W.; Yue, Hao
2015-01-01
This paper proposes a model to address an urban transport planning problem involving combined network design and signal setting in a saturated network. Conventional transport planning models usually deal with the network design problem and signal setting problem separately. However, the fact that network capacity design and capacity allocation determined by network signal setting combine to govern the transport network performance requires the optimal transport planning to consider the two pr...
Wells, Kelley C.; Millet, Dylan B.; Bousserez, Nicolas; Henze, Daven K.; Griffis, Timothy J.; Chaliyakunnel, Sreelekha; Dlugokencky, Edward J.; Saikawa, Eri; Xiang, Gao; Prinn, Ronald G.; O'Doherty, Simon; Young, Dickon; Weiss, Ray F.; Dutton, Geoff S.; Elkins, James W.; Krummel, Paul B.; Langenfelds, Ray; Steele, L. Paul
2018-01-01
We present top-down constraints on global monthly N2O emissions for 2011 from a multi-inversion approach and an ensemble of surface observations. The inversions employ the GEOS-Chem adjoint and an array of aggregation strategies to test how well current observations can constrain the spatial distribution of global N2O emissions. The strategies include (1) a standard 4D-Var inversion at native model resolution (4° × 5°), (2) an inversion for six continental and three ocean regions, and (3) a fast 4D-Var inversion based on a novel dimension reduction technique employing randomized singular value decomposition (SVD). The optimized global flux ranges from 15.9 Tg N yr-1 (SVD-based inversion) to 17.5-17.7 Tg N yr-1 (continental-scale, standard 4D-Var inversions), with the former better capturing the extratropical N2O background measured during the HIAPER Pole-to-Pole Observations (HIPPO) airborne campaigns. We find that the tropics provide a greater contribution to the global N2O flux than is predicted by the prior bottom-up inventories, likely due to underestimated agricultural and oceanic emissions. We infer an overestimate of natural soil emissions in the extratropics and find that predicted emissions are seasonally biased in northern midlatitudes. Here, optimized fluxes exhibit a springtime peak consistent with the timing of spring fertilizer and manure application, soil thawing, and elevated soil moisture. Finally, the inversions reveal a major emission underestimate in the US Corn Belt in the bottom-up inventory used here. We extensively test the impact of initial conditions on the analysis and recommend formally optimizing the initial N2O distribution to avoid biasing the inferred fluxes. We find that the SVD-based approach provides a powerful framework for deriving emission information from N2O observations: by defining the optimal resolution of the solution based on the information content of the inversion, it provides spatial information that is lost when
De Vleeschouwer, David; Da Silva, Anne-Christine; Day, James E.; Whalen, Michael; Claeys, Philippe
2016-04-01
Milankovitch cycles (obliquity, eccentricity and precession) result in changes in the distribution of solar energy over seasons, as well as over latitudes, on time scales of ten thousands of years to millions of years. These changing patterns in insolation have induced significant variations in Earth's past climate over the last 4.5 billion years. Cyclostratigraphy and astrochronology utilize the geologic imprint of such quasi-cyclic climatic variations to measure geologic time. In recent years, major improvements of the Geologic Time Scale have been proposed through the application of cyclostratigraphy, mostly for the Mesozoic and Cenozoic (Gradstein et al., 2012). However, the field of Paleozoic cyclostratigraphy and astrochronology is still in its infancy and the application of cyclostratigraphic techniques in the Paleozoic allows for a whole new range of research questions. For example, unraveling the timing and pacing of environmental changes over the Late Devonian mass extinction on a 105-year time-scale concerns such a novel research question. Here, we present a global cyclostratigraphic framework for late Frasnian to early Famennian climatic and environmental change, through the integration of globally distributed sections. The backbone of this relative time scale consists of previously published cyclostratigraphies for western Canada and Poland (De Vleeschouwer et al., 2012; De Vleeschouwer et al., 2013). We elaborate this Euramerican base by integrating new proxy data -interpreted in terms of astronomical climate forcing- from the Iowa basin (USA, magnetic susceptibility and carbon isotope data) and Belgium (XRF and carbon isotope data). Next, we expand this well-established cyclostratigraphic framework towards the Paleo-Tethys Ocean, using magnetic susceptibility and carbon isotope records from the Fuhe section in South China (Whalen et al., 2015). The resulting global cyclostratigraphic framework implies an important refinement of the late Frasnian to
Energy Technology Data Exchange (ETDEWEB)
Colle, C; Van den Berge, D; De Wagter, C; Fortan, L; Van Duyse, B; De Neve, W
1995-12-01
The design of 3D-conformal dose distributions for targets with concave outlines is a technical challenge in conformal radiotherapy. For these targets, it is impossible to find beam incidences for which the target volume can be isolated from the tissues at risk. Commonly occurring examples are most thyroid cancers and the targets located at the lower neck and upper mediastinal levels related to some head and neck. A solution to this problem was developed, using beam intensity modulation executed with a multileaf collimator by applying a static beam-segmentation technique. The method includes the definition of beam incidences and beam segments of specific shape as well as the calculation of segment weights. Tests on Sherouse`s GRATISTM planning system allowed to escalate the dose to these targets to 65-70 Gy without exceeding spinal cord tolerance. Further optimization by constrained matrix inversion was investigated to explore the possibility of further dose escalation.
Ouyang, Qi; Lu, Wenxi; Hou, Zeyu; Zhang, Yu; Li, Shuai; Luo, Jiannan
2017-05-01
In this paper, a multi-algorithm genetically adaptive multi-objective (AMALGAM) method is proposed as a multi-objective optimization solver. It was implemented in the multi-objective optimization of a groundwater remediation design at sites contaminated by dense non-aqueous phase liquids. In this study, there were two objectives: minimization of the total remediation cost, and minimization of the remediation time. A non-dominated sorting genetic algorithm II (NSGA-II) was adopted to compare with the proposed method. For efficiency, the time-consuming surfactant-enhanced aquifer remediation simulation model was replaced by a surrogate model constructed by a multi-gene genetic programming (MGGP) technique. Similarly, two other surrogate modeling methods-support vector regression (SVR) and Kriging (KRG)-were employed to make comparisons with MGGP. In addition, the surrogate-modeling uncertainty was incorporated in the optimization model by chance-constrained programming (CCP). The results showed that, for the problem considered in this study, (1) the solutions obtained by AMALGAM incurred less remediation cost and required less time than those of NSGA-II, indicating that AMALGAM outperformed NSGA-II. It was additionally shown that (2) the MGGP surrogate model was more accurate than SVR and KRG; and (3) the remediation cost and time increased with the confidence level, which can enable decision makers to make a suitable choice by considering the given budget, remediation time, and reliability. Copyright © 2017 Elsevier B.V. All rights reserved.
Ouyang, Qi; Lu, Wenxi; Hou, Zeyu; Zhang, Yu; Li, Shuai; Luo, Jiannan
2017-05-01
In this paper, a multi-algorithm genetically adaptive multi-objective (AMALGAM) method is proposed as a multi-objective optimization solver. It was implemented in the multi-objective optimization of a groundwater remediation design at sites contaminated by dense non-aqueous phase liquids. In this study, there were two objectives: minimization of the total remediation cost, and minimization of the remediation time. A non-dominated sorting genetic algorithm II (NSGA-II) was adopted to compare with the proposed method. For efficiency, the time-consuming surfactant-enhanced aquifer remediation simulation model was replaced by a surrogate model constructed by a multi-gene genetic programming (MGGP) technique. Similarly, two other surrogate modeling methods-support vector regression (SVR) and Kriging (KRG)-were employed to make comparisons with MGGP. In addition, the surrogate-modeling uncertainty was incorporated in the optimization model by chance-constrained programming (CCP). The results showed that, for the problem considered in this study, (1) the solutions obtained by AMALGAM incurred less remediation cost and required less time than those of NSGA-II, indicating that AMALGAM outperformed NSGA-II. It was additionally shown that (2) the MGGP surrogate model was more accurate than SVR and KRG; and (3) the remediation cost and time increased with the confidence level, which can enable decision makers to make a suitable choice by considering the given budget, remediation time, and reliability.
Bonnet, V; Dumas, R; Cappozzo, A; Joukov, V; Daune, G; Kulić, D; Fraisse, P; Andary, S; Venture, G
2017-09-06
This paper presents a method for real-time estimation of the kinematics and kinetics of a human body performing a sagittal symmetric motor task, which would minimize the impact of the stereophotogrammetric soft tissue artefacts (STA). The method is based on a bi-dimensional mechanical model of the locomotor apparatus the state variables of which (joint angles, velocities and accelerations, and the segments lengths and inertial parameters) are estimated by a constrained extended Kalman filter (CEKF) that fuses input information made of both stereophotogrammetric and dynamometric measurement data. Filter gains are made to saturate in order to obtain plausible state variables and the measurement covariance matrix of the filter accounts for the expected STA maximal amplitudes. We hypothesised that the ensemble of constraints and input redundant information would allow the method to attenuate the STA propagation to the end results. The method was evaluated in ten human subjects performing a squat exercise. The CEKF estimated and measured skin marker trajectories exhibited a RMS difference lower than 4mm, thus in the range of STAs. The RMS differences between the measured ground reaction force and moment and those estimated using the proposed method (9N and 10Nm) were much lower than obtained using a classical inverse dynamics approach (22N and 30Nm). From the latter results it may be inferred that the presented method allows for a significant improvement of the accuracy with which kinematic variables and relevant time derivatives, model parameters and, therefore, intersegmental moments are estimated. Copyright © 2016 Elsevier Ltd. All rights reserved.
Czech Academy of Sciences Publication Activity Database
Branda, Martin; Bucher, M.; Červinka, Michal; Schwartz, A.
2018-01-01
Roč. 70, č. 2 (2018), s. 503-530 ISSN 0926-6003 R&D Projects: GA ČR GA15-00735S Institutional support: RVO:67985556 Keywords : Cardinality constraints * Regularization method * Scholtes regularization * Strong stationarity * Sparse portfolio optimization * Robust portfolio optimization Subject RIV: BB - Applied Statistics, Operational Research OBOR OECD: Statistics and probability Impact factor: 1.520, year: 2016 http://library.utia.cas.cz/separaty/2018/MTR/branda-0489264.pdf
Fast globally optimal segmentation of 3D prostate MRI with axial symmetry prior.
Qiu, Wu; Yuan, Jing; Ukwatta, Eranga; Sun, Yue; Rajchl, Martin; Fenster, Aaron
2013-01-01
We propose a novel global optimization approach to segmenting a given 3D prostate T2w magnetic resonance (MR) image, which enforces the inherent axial symmetry of the prostate shape and simultaneously performs a sequence of 2D axial slice-wise segmentations with a global 3D coherence prior. We show that the proposed challenging combinatorial optimization problem can be solved globally and exactly by means of convex relaxation. With this regard, we introduce a novel coupled continuous max-flow model, which is dual to the studied convex relaxed optimization formulation and leads to an efficient multiplier augmented algorithm based on the modern convex optimization theory. Moreover, the new continuous max-flow based algorithm was implemented on GPUs to achieve a substantial improvement in computation. Experimental results using public and in-house datasets demonstrate great advantages of the proposed method in terms of both accuracy and efficiency.
Global issues and opportunities for optimized retinoblastoma care.
Gallie, Brenda L; Zhao, Junyang; Vandezande, Kirk; White, Abigail; Chan, Helen S L
2007-12-01
The RB1 gene is important in all human cancers. Studies of human retinoblastoma point to a rare retinal cell with extreme dependency on RB1 for initiation but not progression to full malignancy. In developed countries, genetic testing within affected families can predict children at high risk of retinoblastoma before birth; chemotherapy with local therapy often saves eyes and vision; and mortality is 4%. In less developed countries where 92% of children with retinoblastoma are born, mortality reaches 90%. Global collaboration is building for the dramatic change in mortality that awareness, simple expertise and therapies could achieve in less developed countries. Copyright 2007 Wiley-Liss, Inc.
Fast globally optimal segmentation of cells in fluorescence microscopy images.
Bergeest, Jan-Philip; Rohr, Karl
2011-01-01
Accurate and efficient segmentation of cells in fluorescence microscopy images is of central importance for the quantification of protein expression in high-throughput screening applications. We propose a new approach for segmenting cell nuclei which is based on active contours and convex energy functionals. Compared to previous work, our approach determines the global solution. Thus, the approach does not suffer from local minima and the segmentation result does not depend on the initialization. We also suggest a numeric approach for efficiently computing the solution. The performance of our approach has been evaluated using fluorescence microscopy images of different cell types. We have also performed a quantitative comparison with previous segmentation approaches.
A Global Optimization Algorithm for Sum of Linear Ratios Problem
Directory of Open Access Journals (Sweden)
Yuelin Gao
2013-01-01
Full Text Available We equivalently transform the sum of linear ratios programming problem into bilinear programming problem, then by using the linear characteristics of convex envelope and concave envelope of double variables product function, linear relaxation programming of the bilinear programming problem is given, which can determine the lower bound of the optimal value of original problem. Therefore, a branch and bound algorithm for solving sum of linear ratios programming problem is put forward, and the convergence of the algorithm is proved. Numerical experiments are reported to show the effectiveness of the proposed algorithm.
Directory of Open Access Journals (Sweden)
Rica Gonen
2013-11-01
Full Text Available We analyze the space of deterministic, dominant-strategy incentive compatible, individually rational and Pareto optimal combinatorial auctions. We examine a model with multidimensional types, nonidentical items, private values and quasilinear preferences for the players with one relaxation; the players are subject to publicly-known budget constraints. We show that the space includes dictatorial mechanisms and that if dictatorial mechanisms are ruled out by a natural anonymity property, then an impossibility of design is revealed. The same impossibility naturally extends to other abstract mechanisms with an arbitrary outcome set if one maintains the original assumptions of players with quasilinear utilities, public budgets and nonnegative prices.
Hooke–Jeeves Method-used Local Search in a Hybrid Global Optimization Algorithm
Directory of Open Access Journals (Sweden)
V. D. Sulimov
2014-01-01
Full Text Available Modern methods for optimization investigation of complex systems are based on development and updating the mathematical models of systems because of solving the appropriate inverse problems. Input data desirable for solution are obtained from the analysis of experimentally defined consecutive characteristics for a system or a process. Causal characteristics are the sought ones to which equation coefficients of mathematical models of object, limit conditions, etc. belong. The optimization approach is one of the main ones to solve the inverse problems. In the main case it is necessary to find a global extremum of not everywhere differentiable criterion function. Global optimization methods are widely used in problems of identification and computation diagnosis system as well as in optimal control, computing to-mography, image restoration, teaching the neuron networks, other intelligence technologies. Increasingly complicated systems of optimization observed during last decades lead to more complicated mathematical models, thereby making solution of appropriate extreme problems significantly more difficult. A great deal of practical applications may have the problem con-ditions, which can restrict modeling. As a consequence, in inverse problems the criterion functions can be not everywhere differentiable and noisy. Available noise means that calculat-ing the derivatives is difficult and unreliable. It results in using the optimization methods without calculating the derivatives.An efficiency of deterministic algorithms of global optimization is significantly restrict-ed by their dependence on the extreme problem dimension. When the number of variables is large they use the stochastic global optimization algorithms. As stochastic algorithms yield too expensive solutions, so this drawback restricts their applications. Developing hybrid algo-rithms that combine a stochastic algorithm for scanning the variable space with deterministic local search
Yahi, Nouara; Aulas, Anaïs; Fantini, Jacques
2010-02-05
Membrane lipids play a pivotal role in the pathogenesis of Alzheimer's disease, which is associated with conformational changes, oligomerization and/or aggregation of Alzheimer's beta-amyloid (Abeta) peptides. Yet conflicting data have been reported on the respective effect of cholesterol and glycosphingolipids (GSLs) on the supramolecular assembly of Abeta peptides. The aim of the present study was to unravel the molecular mechanisms by which cholesterol modulates the interaction between Abeta(1-40) and chemically defined GSLs (GalCer, LacCer, GM1, GM3). Using the Langmuir monolayer technique, we show that Abeta(1-40) selectively binds to GSLs containing a 2-OH group in the acyl chain of the ceramide backbone (HFA-GSLs). In contrast, Abeta(1-40) did not interact with GSLs containing a nonhydroxylated fatty acid (NFA-GSLs). Cholesterol inhibited the interaction of Abeta(1-40) with HFA-GSLs, through dilution of the GSL in the monolayer, but rendered the initially inactive NFA-GSLs competent for Abeta(1-40) binding. Both crystallographic data and molecular dynamics simulations suggested that the active conformation of HFA-GSL involves a H-bond network that restricts the orientation of the sugar group of GSLs in a parallel orientation with respect to the membrane. This particular conformation is stabilized by the 2-OH group of the GSL. Correspondingly, the interaction of Abeta(1-40) with HFA-GSLs is strongly inhibited by NaF, an efficient competitor of H-bond formation. For NFA-GSLs, this is the OH group of cholesterol that constrains the glycolipid to adopt the active L-shape conformation compatible with sugar-aromatic CH-pi stacking interactions involving residue Y10 of Abeta(1-40). We conclude that cholesterol can either inhibit or facilitate membrane-Abeta interactions through fine tuning of glycosphingolipid conformation. These data shed some light on the complex molecular interplay between cell surface GSLs, cholesterol and Abeta peptides, and on the influence
Directory of Open Access Journals (Sweden)
Nouara Yahi
Full Text Available Membrane lipids play a pivotal role in the pathogenesis of Alzheimer's disease, which is associated with conformational changes, oligomerization and/or aggregation of Alzheimer's beta-amyloid (Abeta peptides. Yet conflicting data have been reported on the respective effect of cholesterol and glycosphingolipids (GSLs on the supramolecular assembly of Abeta peptides. The aim of the present study was to unravel the molecular mechanisms by which cholesterol modulates the interaction between Abeta(1-40 and chemically defined GSLs (GalCer, LacCer, GM1, GM3. Using the Langmuir monolayer technique, we show that Abeta(1-40 selectively binds to GSLs containing a 2-OH group in the acyl chain of the ceramide backbone (HFA-GSLs. In contrast, Abeta(1-40 did not interact with GSLs containing a nonhydroxylated fatty acid (NFA-GSLs. Cholesterol inhibited the interaction of Abeta(1-40 with HFA-GSLs, through dilution of the GSL in the monolayer, but rendered the initially inactive NFA-GSLs competent for Abeta(1-40 binding. Both crystallographic data and molecular dynamics simulations suggested that the active conformation of HFA-GSL involves a H-bond network that restricts the orientation of the sugar group of GSLs in a parallel orientation with respect to the membrane. This particular conformation is stabilized by the 2-OH group of the GSL. Correspondingly, the interaction of Abeta(1-40 with HFA-GSLs is strongly inhibited by NaF, an efficient competitor of H-bond formation. For NFA-GSLs, this is the OH group of cholesterol that constrains the glycolipid to adopt the active L-shape conformation compatible with sugar-aromatic CH-pi stacking interactions involving residue Y10 of Abeta(1-40. We conclude that cholesterol can either inhibit or facilitate membrane-Abeta interactions through fine tuning of glycosphingolipid conformation. These data shed some light on the complex molecular interplay between cell surface GSLs, cholesterol and Abeta peptides, and on the
Directory of Open Access Journals (Sweden)
Joaquin Aranda
2013-08-01
Full Text Available In this paper, we address the problem of determining the optimal geometric configuration of an acoustic sensor network that will maximize the angle-related information available for underwater target positioning. In the set-up adopted, a set of autonomous vehicles carries a network of acoustic units that measure the elevation and azimuth angles between a target and each of the receivers on board the vehicles. It is assumed that the angle measurements are corrupted by white Gaussian noise, the variance of which is distance-dependent. Using tools from estimation theory, the problem is converted into that of minimizing, by proper choice of the sensor positions, the trace of the inverse of the Fisher Information Matrix (also called the Cramer-Rao Bound matrix to determine the sensor configuration that yields the minimum possible covariance of any unbiased target estimator. It is shown that the optimal configuration of the sensors depends explicitly on the intensity of the measurement noise, the constraints imposed on the sensor configuration, the target depth and the probabilistic distribution that defines the prior uncertainty in the target position. Simulation examples illustrate the key results derived.
Moreno-Salinas, David; Pascoal, Antonio; Aranda, Joaquin
2013-08-12
In this paper, we address the problem of determining the optimal geometric configuration of an acoustic sensor network that will maximize the angle-related information available for underwater target positioning. In the set-up adopted, a set of autonomous vehicles carries a network of acoustic units that measure the elevation and azimuth angles between a target and each of the receivers on board the vehicles. It is assumed that the angle measurements are corrupted by white Gaussian noise, the variance of which is distance-dependent. Using tools from estimation theory, the problem is converted into that of minimizing, by proper choice of the sensor positions, the trace of the inverse of the Fisher Information Matrix (also called the Cramer-Rao Bound matrix) to determine the sensor configuration that yields the minimum possible covariance of any unbiased target estimator. It is shown that the optimal configuration of the sensors depends explicitly on the intensity of the measurement noise, the constraints imposed on the sensor configuration, the target depth and the probabilistic distribution that defines the prior uncertainty in the target position. Simulation examples illustrate the key results derived.
Vertical bifacial solar farms: Physics, design, and global optimization
Khan, M. Ryyan
2017-09-04
There have been sustained interest in bifacial solar cell technology since 1980s, with prospects of 30–50% increase in the output power from a stand-alone panel. Moreover, a vertical bifacial panel reduces dust accumulation and provides two output peaks during the day, with the second peak aligned to the peak electricity demand. Recent commercialization and anticipated growth of bifacial panel market have encouraged a closer scrutiny of the integrated power-output and economic viability of bifacial solar farms, where mutual shading will erode some of the anticipated energy gain associated with an isolated, single panel. Towards that goal, in this paper we focus on geography-specific optimization of ground-mounted vertical bifacial solar farms for the entire world. For local irradiance, we combine the measured meteorological data with the clear-sky model. In addition, we consider the effects of direct, diffuse, and albedo light. We assume the panel is configured into sub-strings with bypass-diodes. Based on calculated light collection and panel output, we analyze the optimum farm design for maximum yearly output at any given location in the world. Our results predict that, regardless of the geographical location, a vertical bifacial farm will yield 10–20% more energy than a traditional monofacial farm for a practical row-spacing of 2 m (corresponding to 1.2 m high panels). With the prospect of additional 5–20% energy gain from reduced soiling and tilt optimization, bifacial solar farm do offer a viable technology option for large-scale solar energy generation.
Yang, Dixiong; Liu, Zhenjun; Zhou, Jilei
2014-04-01
Chaos optimization algorithms (COAs) usually utilize the chaotic map like Logistic map to generate the pseudo-random numbers mapped as the design variables for global optimization. Many existing researches indicated that COA can more easily escape from the local minima than classical stochastic optimization algorithms. This paper reveals the inherent mechanism of high efficiency and superior performance of COA, from a new perspective of both the probability distribution property and search speed of chaotic sequences generated by different chaotic maps. The statistical property and search speed of chaotic sequences are represented by the probability density function (PDF) and the Lyapunov exponent, respectively. Meanwhile, the computational performances of hybrid chaos-BFGS algorithms based on eight one-dimensional chaotic maps with different PDF and Lyapunov exponents are compared, in which BFGS is a quasi-Newton method for local optimization. Moreover, several multimodal benchmark examples illustrate that, the probability distribution property and search speed of chaotic sequences from different chaotic maps significantly affect the global searching capability and optimization efficiency of COA. To achieve the high efficiency of COA, it is recommended to adopt the appropriate chaotic map generating the desired chaotic sequences with uniform or nearly uniform probability distribution and large Lyapunov exponent.
Energy Technology Data Exchange (ETDEWEB)
Kooperman, G. J.; Pritchard, M. S.; Ghan, Steven J.; Wang, Minghuai; Somerville, Richard C.; Russell, Lynn
2012-12-11
Natural modes of variability on many timescales influence aerosol particle distributions and cloud properties such that isolating statistically significant differences in cloud radiative forcing due to anthropogenic aerosol perturbations (indirect effects) typically requires integrating over long simulations. For state-of-the-art global climate models (GCM), especially those in which embedded cloud-resolving models replace conventional statistical parameterizations (i.e. multi-scale modeling framework, MMF), the required long integrations can be prohibitively expensive. Here an alternative approach is explored, which implements Newtonian relaxation (nudging) to constrain simulations with both pre-industrial and present-day aerosol emissions toward identical meteorological conditions, thus reducing differences in natural variability and dampening feedback responses in order to isolate radiative forcing. Ten-year GCM simulations with nudging provide a more stable estimate of the global-annual mean aerosol indirect radiative forcing than do conventional free-running simulations. The estimates have mean values and 95% confidence intervals of -1.54 ± 0.02 W/m2 and -1.63 ± 0.17 W/m2 for nudged and free-running simulations, respectively. Nudging also substantially increases the fraction of the world’s area in which a statistically significant aerosol indirect effect can be detected (68% and 25% of the Earth's surface for nudged and free-running simulations, respectively). One-year MMF simulations with and without nudging provide global-annual mean aerosol indirect radiative forcing estimates of -0.80 W/m2 and -0.56 W/m2, respectively. The one-year nudged results compare well with previous estimates from three-year free-running simulations (-0.77 W/m2), which showed the aerosol-cloud relationship to be in better agreement with observations and high-resolution models than in the results obtained with conventional parameterizations.
Raffensperger, Jeff P.; Baker, Anna C.; Blomquist, Joel D.; Hopple, Jessica A.
2017-06-26
Quantitative estimates of base flow are necessary to address questions concerning the vulnerability and response of the Nation’s water supply to natural and human-induced change in environmental conditions. An objective of the U.S. Geological Survey National Water-Quality Assessment Project is to determine how hydrologic systems are affected by watershed characteristics, including land use, land cover, water use, climate, and natural characteristics (geology, soil type, and topography). An important component of any hydrologic system is base flow, generally described as the part of streamflow that is sustained between precipitation events, fed to stream channels by delayed (usually subsurface) pathways, and more specifically as the volumetric discharge of water, estimated at a measurement site or gage at the watershed scale, which represents groundwater that discharges directly or indirectly to stream reaches and is then routed to the measurement point.Hydrograph separation using a recursive digital filter was applied to 225 sites in the Chesapeake Bay watershed. The recursive digital filter was chosen for the following reasons: it is based in part on the assumption that groundwater acts as a linear reservoir, and so has a physical basis; it has only two adjustable parameters (alpha, obtained directly from recession analysis, and beta, the maximum value of the base-flow index that can be modeled by the filter), which can be determined objectively and with the same physical basis of groundwater reservoir linearity, or that can be optimized by applying a chemical-mass-balance constraint. Base-flow estimates from the recursive digital filter were compared with those from five other hydrograph-separation methods with respect to two metrics: the long-term average fraction of streamflow that is base flow, or base-flow index, and the fraction of days where streamflow is entirely base flow. There was generally good correlation between the methods, with some biased
DEFF Research Database (Denmark)
Mirzaei, Mahmood; Tibaldi, Carlo; Hansen, Morten Hartvig
2016-01-01
to automatically tune the wind turbine controller subject to robustness constraints. The properties of the system such as the maximum sensitivity and complementary sensitivity functions (Ms and Mt), along with some of the responses of the system, are used to investigate the controller performance and formulate...... the optimization problem. The cost function is the integral absolute error (IAE) of the rotational speed from a disturbance modeled as a step in wind speed. Linearized model of the DTU 10-MW reference wind turbine is obtained using HAWCStab2. Thereafter, the model is reduced with model order reduction. The trade......PI/PID controllers are the most common wind turbine controllers. Normally a first tuning is obtained using methods such as pole-placement or Ziegler-Nichols and then extensive aeroelastic simulations are used to obtain the best tuning in terms of regulation of the outputs and reduction of the loads...
A global optimization method for evaporative cooling systems based on the entransy theory
International Nuclear Information System (INIS)
Yuan, Fang; Chen, Qun
2012-01-01
Evaporative cooling technique, one of the most widely used methods, is essential to both energy conservation and environment protection. This contribution introduces a global optimization method for indirect evaporative cooling systems with coupled heat and mass transfer processes based on the entransy theory to improve their energy efficiency. First, we classify the irreversible processes in the system into the heat transfer process, the coupled heat and mass transfer process and the mixing process of waters in different branches, where the irreversibility is evaluated by the entransy dissipation. Then through the total system entransy dissipation, we establish the theoretical relationship of the user demands with both the geometrical structures of each heat exchanger and the operating parameters of each fluid, and derive two optimization equation groups focusing on two typical optimization problems. Finally, an indirect evaporative cooling system is taken as an example to illustrate the applications of the newly proposed optimization method. It is concluded that there exists an optimal circulating water flow rate with the minimum total thermal conductance of the system. Furthermore, with different user demands and moist air inlet conditions, it is the global optimization, other than parametric analysis, will obtain the optimal performance of the system. -- Highlights: ► Introduce a global optimization method for evaporative cooling systems. ► Establish the direct relation between user demands and the design parameters. ► Obtain two groups of optimization equations for two typical optimization objectives. ► Solving the equations offers the optimal design parameters for the system. ► Provide the instruction for the design of coupled heat and mass transfer systems.
International Nuclear Information System (INIS)
Xu, Yun-Chao; Chen, Qun
2013-01-01
The vapor-compression refrigeration systems have been one of the essential energy conversion systems for humankind and exhausting huge amounts of energy nowadays. Surrounding the energy efficiency promotion of the systems, there are lots of effectual optimization methods but mainly relied on engineering experience and computer simulations rather than theoretical analysis due to the complex and vague physical essence. We attempt to propose a theoretical global optimization method based on in-depth physical analysis for the involved physical processes, i.e. heat transfer analysis for condenser and evaporator, through introducing the entransy theory and thermodynamic analysis for compressor and expansion valve. The integration of heat transfer and thermodynamic analyses forms the overall physical optimization model for the systems to describe the relation between all the unknown parameters and known conditions, which makes theoretical global optimization possible. With the aid of the mathematical conditional extremum solutions, an optimization equation group and the optimal configuration of all the unknown parameters are analytically obtained. Eventually, via the optimization of a typical vapor-compression refrigeration system with various working conditions to minimize the total heat transfer area of heat exchangers, the validity and superior of the newly proposed optimization method is proved. - Highlights: • A global optimization method for vapor-compression systems is proposed. • Integrating heat transfer and thermodynamic analyses forms the optimization model. • A mathematical relation between design parameters and requirements is derived. • Entransy dissipation is introduced into heat transfer analysis. • The validity of the method is proved via optimization of practical cases
Globally optimal, minimum stored energy, double-doughnut superconducting magnets.
Tieng, Quang M; Vegh, Viktor; Brereton, Ian M
2010-01-01
The use of the minimum stored energy current density map-based methodology of designing closed-bore symmetric superconducting magnets was described recently. The technique is further developed to cater for the design of interventional-type MRI systems, and in particular open symmetric magnets of the double-doughnut configuration. This extends the work to multiple magnet domain configurations. The use of double-doughnut magnets in MRI scanners has previously been hindered by the ability to deliver strong magnetic fields over a sufficiently large volume appropriate for imaging, essentially limiting spatial resolution, signal-to-noise ratio, and field of view. The requirement of dedicated interventional space restricts the manner in which the coils can be arranged and placed. The minimum stored energy optimal coil arrangement ensures that the field strength is maximized over a specific region of imaging. The design method yields open, dual-domain magnets capable of delivering greater field strengths than those used prior to this work, and at the same time it provides an increase in the field-of-view volume. Simulation results are provided for 1-T double-doughnut magnets with at least a 50-cm 1-ppm (parts per million) field of view and 0.7-m gap between the two doughnuts. Copyright (c) 2009 Wiley-Liss, Inc.
International Nuclear Information System (INIS)
Poursalehi, N.; Zolfaghari, A.; Minuchehr, A.; Valavi, K.
2013-01-01
Highlights: • SGHS enhanced the convergence rate of LPO using some improvements in comparison to basic HS and GHS. • SGHS optimization algorithm obtained averagely better fitness relative to basic HS and GHS algorithms. • Upshot of the SGHS implementation in the LPO reveals its flexibility, efficiency and reliability. - Abstract: The aim of this work is to apply the new developed optimization algorithm, Self-adaptive Global best Harmony Search (SGHS), for PWRs fuel management optimization. SGHS algorithm has some modifications in comparison with basic Harmony Search (HS) and Global-best Harmony Search (GHS) algorithms such as dynamically change of parameters. For the demonstration of SGHS ability to find an optimal configuration of fuel assemblies, basic Harmony Search (HS) and Global-best Harmony Search (GHS) algorithms also have been developed and investigated. For this purpose, Self-adaptive Global best Harmony Search Nodal Expansion package (SGHSNE) has been developed implementing HS, GHS and SGHS optimization algorithms for the fuel management operation of nuclear reactor cores. This package uses developed average current nodal expansion code which solves the multi group diffusion equation by employment of first and second orders of Nodal Expansion Method (NEM) for two dimensional, hexagonal and rectangular geometries, respectively, by one node per a FA. Loading pattern optimization was performed using SGHSNE package for some test cases to present the SGHS algorithm capability in converging to near optimal loading pattern. Results indicate that the convergence rate and reliability of the SGHS method are quite promising and practically, SGHS improves the quality of loading pattern optimization results relative to HS and GHS algorithms. As a result, it has the potential to be used in the other nuclear engineering optimization problems
A Guiding Evolutionary Algorithm with Greedy Strategy for Global Optimization Problems
Directory of Open Access Journals (Sweden)
Leilei Cao
2016-01-01
Full Text Available A Guiding Evolutionary Algorithm (GEA with greedy strategy for global optimization problems is proposed. Inspired by Particle Swarm Optimization, the Genetic Algorithm, and the Bat Algorithm, the GEA was designed to retain some advantages of each method while avoiding some disadvantages. In contrast to the usual Genetic Algorithm, each individual in GEA is crossed with the current global best one instead of a randomly selected individual. The current best individual served as a guide to attract offspring to its region of genotype space. Mutation was added to offspring according to a dynamic mutation probability. To increase the capability of exploitation, a local search mechanism was applied to new individuals according to a dynamic probability of local search. Experimental results show that GEA outperformed the other three typical global optimization algorithms with which it was compared.
The Global Optimal Algorithm of Reliable Path Finding Problem Based on Backtracking Method
Directory of Open Access Journals (Sweden)
Liang Shen
2017-01-01
Full Text Available There is a growing interest in finding a global optimal path in transportation networks particularly when the network suffers from unexpected disturbance. This paper studies the problem of finding a global optimal path to guarantee a given probability of arriving on time in a network with uncertainty, in which the travel time is stochastic instead of deterministic. Traditional path finding methods based on least expected travel time cannot capture the network user’s risk-taking behaviors in path finding. To overcome such limitation, the reliable path finding algorithms have been proposed but the convergence of global optimum is seldom addressed in the literature. This paper integrates the K-shortest path algorithm into Backtracking method to propose a new path finding algorithm under uncertainty. The global optimum of the proposed method can be guaranteed. Numerical examples are conducted to demonstrate the correctness and efficiency of the proposed algorithm.
Conference on "State of the Art in Global Optimization : Computational Methods and Applications"
Pardalos, P
1996-01-01
Optimization problems abound in most fields of science, engineering, and technology. In many of these problems it is necessary to compute the global optimum (or a good approximation) of a multivariable function. The variables that define the function to be optimized can be continuous and/or discrete and, in addition, many times satisfy certain constraints. Global optimization problems belong to the complexity class of NP-hard prob lems. Such problems are very difficult to solve. Traditional descent optimization algorithms based on local information are not adequate for solving these problems. In most cases of practical interest the number of local optima increases, on the aver age, exponentially with the size of the problem (number of variables). Furthermore, most of the traditional approaches fail to escape from a local optimum in order to continue the search for the global solution. Global optimization has received a lot of attention in the past ten years, due to the success of new algorithms for solvin...
Global warming and carbon taxation. Optimal policy and the role of administration costs
International Nuclear Information System (INIS)
Williams, M.
1995-01-01
This paper develops a model relating CO 2 emissions to atmosphere concentrations, global temperature change and economic damages. For a variety of parameter assumptions, the model provides estimates of the marginal cost of emissions in various years. The optimal carbon tax is a function of the marginal emission cost and the costs of administering the tax. This paper demonstrates that under any reasonable assumptions, the optimal carbon tax is zero for at least several decades. (author)
International Nuclear Information System (INIS)
Dong, Huachao; Song, Baowei; Wang, Peng; Huang, Shuai
2015-01-01
In this paper, a novel kriging-based algorithm for global optimization of computationally expensive black-box functions is presented. This algorithm utilizes a multi-start approach to find all of the local optimal values of the surrogate model and performs searches within the neighboring area around these local optimal positions. Compared with traditional surrogate-based global optimization method, this algorithm provides another kind of balance between exploitation and exploration on kriging-based model. In addition, a new search strategy is proposed and coupled into this optimization process. The local search strategy employs a kind of improved 'Minimizing the predictor' method, which dynamically adjusts search direction and radius until finds the optimal value. Furthermore, the global search strategy utilizes the advantage of kriging-based model in predicting unexplored regions to guarantee the reliability of the algorithm. Finally, experiments on 13 test functions with six algorithms are set up and the results show that the proposed algorithm is very promising.
Optimal interpolation schemes to constrain pmPM2.5 in regional modeling over the United States
Sousan, Sinan Dhia Jameel
This thesis presents the use of data assimilation with optimal interpolation (OI) to develop atmospheric aerosol concentration estimates for the United States at high spatial and temporal resolutions. Concentration estimates are highly desirable for a wide range of applications, including visibility, climate, and human health. OI is a viable data assimilation method that can be used to improve Community Multiscale Air Quality (CMAQ) model fine particulate matter (PM2.5) estimates. PM2.5 is the mass of solid and liquid particles with diameters less than or equal to 2.5 µm suspended in the gas phase. OI was employed by combining model estimates with satellite and surface measurements. The satellite data assimilation combined 36 x 36 km aerosol concentrations from CMAQ with aerosol optical depth (AOD) measured by MODIS and AERONET over the continental United States for 2002. Posterior model concentrations generated by the OI algorithm were compared with surface PM2.5 measurements to evaluate a number of possible data assimilation parameters, including model error, observation error, and temporal averaging assumptions. Evaluation was conducted separately for six geographic U.S. regions in 2002. Variability in model error and MODIS biases limited the effectiveness of a single data assimilation system for the entire continental domain. The best combinations of four settings and three averaging schemes led to a domain-averaged improvement in fractional error from 1.2 to 0.97 and from 0.99 to 0.89 at respective IMPROVE and STN monitoring sites. For 38% of OI results, MODIS OI degraded the forward model skill due to biases and outliers in MODIS AOD. Surface data assimilation combined 36 × 36 km aerosol concentrations from the CMAQ model with surface PM2.5 measurements over the continental United States for 2002. The model error covariance matrix was constructed by using the observational method. The observation error covariance matrix included site representation that
Theoretical properties of the global optimizer of two layer neural network
Boob, Digvijay; Lan, Guanghui
2017-01-01
In this paper, we study the problem of optimizing a two-layer artificial neural network that best fits a training dataset. We look at this problem in the setting where the number of parameters is greater than the number of sampled points. We show that for a wide class of differentiable activation functions (this class involves "almost" all functions which are not piecewise linear), we have that first-order optimal solutions satisfy global optimality provided the hidden layer is non-singular. ...
A global optimization algorithm inspired in the behavior of selfish herds.
Fausto, Fernando; Cuevas, Erik; Valdivia, Arturo; González, Adrián
2017-10-01
In this paper, a novel swarm optimization algorithm called the Selfish Herd Optimizer (SHO) is proposed for solving global optimization problems. SHO is based on the simulation of the widely observed selfish herd behavior manifested by individuals within a herd of animals subjected to some form of predation risk. In SHO, individuals emulate the predatory interactions between groups of prey and predators by two types of search agents: the members of a selfish herd (the prey) and a pack of hungry predators. Depending on their classification as either a prey or a predator, each individual is conducted by a set of unique evolutionary operators inspired by such prey-predator relationship. These unique traits allow SHO to improve the balance between exploration and exploitation without altering the population size. To illustrate the proficiency and robustness of the proposed method, it is compared to other well-known evolutionary optimization approaches such as Particle Swarm Optimization (PSO), Artificial Bee Colony (ABC), Firefly Algorithm (FA), Differential Evolution (DE), Genetic Algorithms (GA), Crow Search Algorithm (CSA), Dragonfly Algorithm (DA), Moth-flame Optimization Algorithm (MOA) and Sine Cosine Algorithm (SCA). The comparison examines several standard benchmark functions, commonly considered within the literature of evolutionary algorithms. The experimental results show the remarkable performance of our proposed approach against those of the other compared methods, and as such SHO is proven to be an excellent alternative to solve global optimization problems. Copyright © 2017 Elsevier B.V. All rights reserved.
Global optimal path planning of an autonomous vehicle for overtaking a moving obstacle
Directory of Open Access Journals (Sweden)
B. Mashadi
Full Text Available In this paper, the global optimal path planning of an autonomous vehicle for overtaking a moving obstacle is proposed. In this study, the autonomous vehicle overtakes a moving vehicle by performing a double lane-change maneuver after detecting it in a proper distance ahead. The optimal path of vehicle for performing the lane-change maneuver is generated by a path planning program in which the sum of lateral deviation of the vehicle from a reference path and the rate of steering angle become minimum while the lateral acceleration of vehicle does not exceed a safe limit value. A nonlinear optimal control theory with the lateral vehicle dynamics equations and inequality constraint of lateral acceleration are used to generate the path. The indirect approach for solving the optimal control problem is used by applying the calculus of variation and the Pontryagin's Minimum Principle to obtain first-order necessary conditions for optimality. The optimal path is generated as a global optimal solution and can be used as the benchmark of the path generated by the local motion planning of autonomous vehicles. A full nonlinear vehicle model in CarSim software is used for path following simulation by importing path data from the MATLAB code. The simulation results show that the generated path for the autonomous vehicle satisfies all vehicle dynamics constraints and hence is a suitable overtaking path for the following vehicle.
CSIR Research Space (South Africa)
Britz, K
2011-09-01
Full Text Available their basic properties and relationship. In Section 3 we present a modal instance of these constructions which also illustrates with an example how to reason abductively with constrained entailment in a causal or action oriented context. In Section 4 we... of models with the former approach, whereas in Section 3.3 we give an example illustrating ways in which C can be de ned with both. Here we employ the following versions of local consequence: De nition 3.4. Given a model M = hW;R;Vi and formulas...
Liang, Faming; Cheng, Yichen; Lin, Guang
2014-01-01
cooling schedule, for example, a square-root cooling schedule, while guaranteeing the global optima to be reached when the temperature tends to zero. The new algorithm has been tested on a few benchmark optimization problems, including feed-forward neural
PS-FW: A Hybrid Algorithm Based on Particle Swarm and Fireworks for Global Optimization
Chen, Shuangqing; Wei, Lixin; Guan, Bing
2018-01-01
Particle swarm optimization (PSO) and fireworks algorithm (FWA) are two recently developed optimization methods which have been applied in various areas due to their simplicity and efficiency. However, when being applied to high-dimensional optimization problems, PSO algorithm may be trapped in the local optima owing to the lack of powerful global exploration capability, and fireworks algorithm is difficult to converge in some cases because of its relatively low local exploitation efficiency for noncore fireworks. In this paper, a hybrid algorithm called PS-FW is presented, in which the modified operators of FWA are embedded into the solving process of PSO. In the iteration process, the abandonment and supplement mechanism is adopted to balance the exploration and exploitation ability of PS-FW, and the modified explosion operator and the novel mutation operator are proposed to speed up the global convergence and to avoid prematurity. To verify the performance of the proposed PS-FW algorithm, 22 high-dimensional benchmark functions have been employed, and it is compared with PSO, FWA, stdPSO, CPSO, CLPSO, FIPS, Frankenstein, and ALWPSO algorithms. Results show that the PS-FW algorithm is an efficient, robust, and fast converging optimization method for solving global optimization problems. PMID:29675036
Energy Technology Data Exchange (ETDEWEB)
Almeida Bezerra, Marcos, E-mail: mbezerra47@yahoo.com.br [Universidade Estadual do Sudoeste da Bahia, Laboratorio de Quimica Analitica, 45200-190, Jequie, Bahia (Brazil); Teixeira Castro, Jacira [Universidade Federal do Reconcavo da Bahia, Centro de Ciencias Exatas e Tecnologicas, 44380-000, Cruz das Almas, Bahia (Brazil); Coelho Macedo, Reinaldo; Goncalves da Silva, Douglas [Universidade Estadual do Sudoeste da Bahia, Laboratorio de Quimica Analitica, 45200-190, Jequie, Bahia (Brazil)
2010-06-18
A slurry suspension sampling technique has been developed for manganese and zinc determination in tea leaves by using flame atomic absorption spectrometry. The proportions of liquid-phase of the slurries composed by HCl, HNO{sub 3} and Triton X-100 solutions have been optimized applying a constrained mixture design. The optimized conditions were 200 mg of sample ground in a tungsten carbide balls mill (particle size < 100 {mu}m), dilution in a liquid-phase composed by 2.0 mol L{sup -1} nitric, 2.0 mol L{sup -1} hydrochloric acid and 2.5% Triton X-100 solutions (in the proportions of 50%, 12% and 38% respectively), sonication time of 10 min and final slurry volume of 50.0 mL. This method allowed the determination of manganese and zinc by FAAS, with detection limits of 0.46 and 0.66 {mu}g g{sup -1}, respectively. The precisions, expressed as relative standard deviation (RSD), are 6.9 and 5.5% (n = 10), for concentrations of manganese and zinc of 20 and 40 {mu}g g{sup -1}, respectively. The accuracy of the method was confirmed by analysis of the certified apple leaves (NIST 1515) and spinach leaves (NIST 1570a). The proposed method was applied for the determination of manganese and zinc in tea leaves used for the preparation of infusions. The obtained concentrations varied between 42 and 118 {mu}g g{sup -1} and 18.6 and 90 {mu}g g{sup -1}, respectively, for manganese and zinc. The results were compared with those obtained by an acid digestion procedure and determination of the elements by FAAS. There was no significant difference between the results obtained by the two methods based on a paired t-test (at 95% confidence level).
Selective Segmentation for Global Optimization of Depth Estimation in Complex Scenes
Directory of Open Access Journals (Sweden)
Sheng Liu
2013-01-01
Full Text Available This paper proposes a segmentation-based global optimization method for depth estimation. Firstly, for obtaining accurate matching cost, the original local stereo matching approach based on self-adapting matching window is integrated with two matching cost optimization strategies aiming at handling both borders and occlusion regions. Secondly, we employ a comprehensive smooth term to satisfy diverse smoothness request in real scene. Thirdly, a selective segmentation term is used for enforcing the plane trend constraints selectively on the corresponding segments to further improve the accuracy of depth results from object level. Experiments on the Middlebury image pairs show that the proposed global optimization approach is considerably competitive with other state-of-the-art matching approaches.
Global optimization based on noisy evaluations: An empirical study of two statistical approaches
International Nuclear Information System (INIS)
Vazquez, Emmanuel; Villemonteix, Julien; Sidorkiewicz, Maryan; Walter, Eric
2008-01-01
The optimization of the output of complex computer codes has often to be achieved with a small budget of evaluations. Algorithms dedicated to such problems have been developed and compared, such as the Expected Improvement algorithm (El) or the Informational Approach to Global Optimization (IAGO). However, the influence of noisy evaluation results on the outcome of these comparisons has often been neglected, despite its frequent appearance in industrial problems. In this paper, empirical convergence rates for El and IAGO are compared when an additive noise corrupts the result of an evaluation. IAGO appears more efficient than El and various modifications of El designed to deal with noisy evaluations. Keywords. Global optimization; computer simulations; kriging; Gaussian process; noisy evaluations.
Huang, Si-Da; Shang, Cheng; Zhang, Xiao-Jie; Liu, Zhi-Pan
2017-09-01
While the underlying potential energy surface (PES) determines the structure and other properties of a material, it has been frustrating to predict new materials from theory even with the advent of supercomputing facilities. The accuracy of the PES and the efficiency of PES sampling are two major bottlenecks, not least because of the great complexity of the material PES. This work introduces a "Global-to-Global" approach for material discovery by combining for the first time a global optimization method with neural network (NN) techniques. The novel global optimization method, named the stochastic surface walking (SSW) method, is carried out massively in parallel for generating a global training data set, the fitting of which by the atom-centered NN produces a multi-dimensional global PES; the subsequent SSW exploration of large systems with the analytical NN PES can provide key information on the thermodynamics and kinetics stability of unknown phases identified from global PESs. We describe in detail the current implementation of the SSW-NN method with particular focuses on the size of the global data set and the simultaneous energy/force/stress NN training procedure. An important functional material, TiO 2 , is utilized as an example to demonstrate the automated global data set generation, the improved NN training procedure and the application in material discovery. Two new TiO 2 porous crystal structures are identified, which have similar thermodynamics stability to the common TiO 2 rutile phase and the kinetics stability for one of them is further proved from SSW pathway sampling. As a general tool for material simulation, the SSW-NN method provides an efficient and predictive platform for large-scale computational material screening.
Ariyarit, Atthaphon; Sugiura, Masahiko; Tanabe, Yasutada; Kanazaki, Masahiro
2018-06-01
A multi-fidelity optimization technique by an efficient global optimization process using a hybrid surrogate model is investigated for solving real-world design problems. The model constructs the local deviation using the kriging method and the global model using a radial basis function. The expected improvement is computed to decide additional samples that can improve the model. The approach was first investigated by solving mathematical test problems. The results were compared with optimization results from an ordinary kriging method and a co-kriging method, and the proposed method produced the best solution. The proposed method was also applied to aerodynamic design optimization of helicopter blades to obtain the maximum blade efficiency. The optimal shape obtained by the proposed method achieved performance almost equivalent to that obtained using the high-fidelity, evaluation-based single-fidelity optimization. Comparing all three methods, the proposed method required the lowest total number of high-fidelity evaluation runs to obtain a converged solution.
Calibration of an agricultural-hydrological model (RZWQM2) using surrogate global optimization
Xi, Maolong; Lu, Dan; Gui, Dongwei; Qi, Zhiming; Zhang, Guannan
2017-01-01
Robust calibration of an agricultural-hydrological model is critical for simulating crop yield and water quality and making reasonable agricultural management. However, calibration of the agricultural-hydrological system models is challenging because of model complexity, the existence of strong parameter correlation, and significant computational requirements. Therefore, only a limited number of simulations can be allowed in any attempt to find a near-optimal solution within an affordable time, which greatly restricts the successful application of the model. The goal of this study is to locate the optimal solution of the Root Zone Water Quality Model (RZWQM2) given a limited simulation time, so as to improve the model simulation and help make rational and effective agricultural-hydrological decisions. To this end, we propose a computationally efficient global optimization procedure using sparse-grid based surrogates. We first used advanced sparse grid (SG) interpolation to construct a surrogate system of the actual RZWQM2, and then we calibrate the surrogate model using the global optimization algorithm, Quantum-behaved Particle Swarm Optimization (QPSO). As the surrogate model is a polynomial with fast evaluation, it can be efficiently evaluated with a sufficiently large number of times during the optimization, which facilitates the global search. We calibrate seven model parameters against five years of yield, drain flow, and NO3-N loss data from a subsurface-drained corn-soybean field in Iowa. Results indicate that an accurate surrogate model can be created for the RZWQM2 with a relatively small number of SG points (i.e., RZWQM2 runs). Compared to the conventional QPSO algorithm, our surrogate-based optimization method can achieve a smaller objective function value and better calibration performance using a fewer number of expensive RZWQM2 executions, which greatly improves computational efficiency.
Statistical distributions of optimal global alignment scores of random protein sequences
Directory of Open Access Journals (Sweden)
Tang Jiaowei
2005-10-01
Full Text Available Abstract Background The inference of homology from statistically significant sequence similarity is a central issue in sequence alignments. So far the statistical distribution function underlying the optimal global alignments has not been completely determined. Results In this study, random and real but unrelated sequences prepared in six different ways were selected as reference datasets to obtain their respective statistical distributions of global alignment scores. All alignments were carried out with the Needleman-Wunsch algorithm and optimal scores were fitted to the Gumbel, normal and gamma distributions respectively. The three-parameter gamma distribution performs the best as the theoretical distribution function of global alignment scores, as it agrees perfectly well with the distribution of alignment scores. The normal distribution also agrees well with the score distribution frequencies when the shape parameter of the gamma distribution is sufficiently large, for this is the scenario when the normal distribution can be viewed as an approximation of the gamma distribution. Conclusion We have shown that the optimal global alignment scores of random protein sequences fit the three-parameter gamma distribution function. This would be useful for the inference of homology between sequences whose relationship is unknown, through the evaluation of gamma distribution significance between sequences.
Directory of Open Access Journals (Sweden)
Weitian Lin
2014-01-01
Full Text Available Particle swarm optimization algorithm (PSOA is an advantage optimization tool. However, it has a tendency to get stuck in a near optimal solution especially for middle and large size problems and it is difficult to improve solution accuracy by fine-tuning parameters. According to the insufficiency, this paper researches the local and global search combine particle swarm algorithm (LGSCPSOA, and its convergence and obtains its convergence qualification. At the same time, it is tested with a set of 8 benchmark continuous functions and compared their optimization results with original particle swarm algorithm (OPSOA. Experimental results indicate that the LGSCPSOA improves the search performance especially on the middle and large size benchmark functions significantly.
Energy Technology Data Exchange (ETDEWEB)
Cho, Su Gil; Jang, Jun Yong; Kim, Ji Hoon; Lee, Tae Hee [Hanyang University, Seoul (Korea, Republic of); Lee, Min Uk [Romax Technology Ltd., Seoul (Korea, Republic of); Choi, Jong Su; Hong, Sup [Korea Research Institute of Ships and Ocean Engineering, Daejeon (Korea, Republic of)
2015-04-15
Sequential surrogate model-based global optimization algorithms, such as super-EGO, have been developed to increase the efficiency of commonly used global optimization technique as well as to ensure the accuracy of optimization. However, earlier studies have drawbacks because there are three phases in the optimization loop and empirical parameters. We propose a united sampling criterion to simplify the algorithm and to achieve the global optimum of problems with constraints without any empirical parameters. It is able to select the points located in a feasible region with high model uncertainty as well as the points along the boundary of constraint at the lowest objective value. The mean squared error determines which criterion is more dominant among the infill sampling criterion and boundary sampling criterion. Also, the method guarantees the accuracy of the surrogate model because the sample points are not located within extremely small regions like super-EGO. The performance of the proposed method, such as the solvability of a problem, convergence properties, and efficiency, are validated through nonlinear numerical examples with disconnected feasible regions.
Energy Technology Data Exchange (ETDEWEB)
Rattá, G.A., E-mail: giuseppe.ratta@ciemat.es [Laboratorio Nacional de Fusión, CIEMAT, Madrid (Spain); Vega, J. [Laboratorio Nacional de Fusión, CIEMAT, Madrid (Spain); Murari, A. [Consorzio RFX, Associazione EURATOM/ENEA per la Fusione, Padua (Italy); Dormido-Canto, S. [Dpto. de Informática y Automática, Universidad Nacional de Educación a Distancia, Madrid (Spain); Moreno, R. [Laboratorio Nacional de Fusión, CIEMAT, Madrid (Spain)
2016-11-15
Highlights: • A global optimization method based on genetic algorithms was developed. • It allowed improving the prediction of disruptions using APODIS architecture. • It also provides the potential opportunity to develop a spectrum of future predictors using different training datasets. • The future analysis of how their structures reassemble and evolve in each test may help to improve the development of disruption predictors for ITER. - Abstract: Since year 2010, the APODIS architecture has proven its accuracy predicting disruptions in JET tokamak. Nevertheless, it has shown margins for improvements, fact indisputable after the enhanced performances achieved in posterior upgrades. In this article, a complete optimization driven by Genetic Algorithms (GA) is applied to it aiming at considering all possible combination of signals, signal features, quantity of models, their characteristics and internal parameters. This global optimization targets the creation of the best possible system with a reduced amount of required training data. The results harbor no doubts about the reliability of the global optimization method, allowing to outperform the ones of previous versions: 91.77% of predictions (89.24% with an anticipation higher than 10 ms) with a 3.55% of false alarms. Beyond its effectiveness, it also provides the potential opportunity to develop a spectrum of future predictors using different training datasets.
International Nuclear Information System (INIS)
Rattá, G.A.; Vega, J.; Murari, A.; Dormido-Canto, S.; Moreno, R.
2016-01-01
Highlights: • A global optimization method based on genetic algorithms was developed. • It allowed improving the prediction of disruptions using APODIS architecture. • It also provides the potential opportunity to develop a spectrum of future predictors using different training datasets. • The future analysis of how their structures reassemble and evolve in each test may help to improve the development of disruption predictors for ITER. - Abstract: Since year 2010, the APODIS architecture has proven its accuracy predicting disruptions in JET tokamak. Nevertheless, it has shown margins for improvements, fact indisputable after the enhanced performances achieved in posterior upgrades. In this article, a complete optimization driven by Genetic Algorithms (GA) is applied to it aiming at considering all possible combination of signals, signal features, quantity of models, their characteristics and internal parameters. This global optimization targets the creation of the best possible system with a reduced amount of required training data. The results harbor no doubts about the reliability of the global optimization method, allowing to outperform the ones of previous versions: 91.77% of predictions (89.24% with an anticipation higher than 10 ms) with a 3.55% of false alarms. Beyond its effectiveness, it also provides the potential opportunity to develop a spectrum of future predictors using different training datasets.
International Nuclear Information System (INIS)
Zhou Jinghao; Kim, Sung; Jabbour, Salma; Goyal, Sharad; Haffty, Bruce; Chen, Ting; Levinson, Lydia; Metaxas, Dimitris; Yue, Ning J.
2010-01-01
Purpose: In the external beam radiation treatment of prostate cancers, successful implementation of adaptive radiotherapy and conformal radiation dose delivery is highly dependent on precise and expeditious segmentation and registration of the prostate volume between the simulation and the treatment images. The purpose of this study is to develop a novel, fast, and accurate segmentation and registration method to increase the computational efficiency to meet the restricted clinical treatment time requirement in image guided radiotherapy. Methods: The method developed in this study used soft tissues to capture the transformation between the 3D planning CT (pCT) images and 3D cone-beam CT (CBCT) treatment images. The method incorporated a global-to-local deformable mesh model based registration framework as well as an automatic anatomy-constrained robust active shape model (ACRASM) based segmentation algorithm in the 3D CBCT images. The global registration was based on the mutual information method, and the local registration was to minimize the Euclidian distance of the corresponding nodal points from the global transformation of deformable mesh models, which implicitly used the information of the segmented target volume. The method was applied on six data sets of prostate cancer patients. Target volumes delineated by the same radiation oncologist on the pCT and CBCT were chosen as the benchmarks and were compared to the segmented and registered results. The distance-based and the volume-based estimators were used to quantitatively evaluate the results of segmentation and registration. Results: The ACRASM segmentation algorithm was compared to the original active shape model (ASM) algorithm by evaluating the values of the distance-based estimators. With respect to the corresponding benchmarks, the mean distance ranged from -0.85 to 0.84 mm for ACRASM and from -1.44 to 1.17 mm for ASM. The mean absolute distance ranged from 1.77 to 3.07 mm for ACRASM and from 2.45 to
International Nuclear Information System (INIS)
Zou, Dexuan; Li, Steven; Li, Zongyan; Kong, Xiangyong
2017-01-01
Highlights: • A new global particle swarm optimization (NGPSO) is proposed. • NGPSO has strong convergence and desirable accuracy. • NGPSO is used to handle the economic emission dispatch with or without transmission losses. • The equality constraint can be satisfied by solving a quadratic equation. • The inequality constraints can be satisfied by using penalty function method. - Abstract: A new global particle swarm optimization (NGPSO) algorithm is proposed to solve the economic emission dispatch (EED) problems in this paper. NGPSO is different from the traditional particle swarm optimization (PSO) algorithm in two aspects. First, NGPSO uses a new position updating equation which relies on the global best particle to guide the searching activities of all particles. Second, it uses the randomization based on the uniform distribution to slightly disturb the flight trajectories of particles during the late evolutionary process. The two steps enable NGPSO to effectively execute a number of global searches, and thus they increase the chance of exploring promising solution space, and reduce the probabilities of getting trapped into local optima for all particles. On the other hand, the two objective functions of EED are normalized separately according to all candidate solutions, and then they are incorporated into one single objective function. The transformation steps are very helpful in eliminating the difference caused by the different dimensions of the two functions, and thus they strike a balance between the fuel cost and emission. In addition, a simple and common penalty function method is employed to facilitate the satisfactions of EED’s constraints. Based on these improvements in PSO, objective functions and constraints handling, high-quality solutions can be obtained for EED problems. Five examples are chosen to testify the performance of three improved PSOs on solving EED problems with or without transmission losses. Experimental results show that
SGO: A fast engine for ab initio atomic structure global optimization by differential evolution
Chen, Zhanghui; Jia, Weile; Jiang, Xiangwei; Li, Shu-Shen; Wang, Lin-Wang
2017-10-01
As the high throughout calculations and material genome approaches become more and more popular in material science, the search for optimal ways to predict atomic global minimum structure is a high research priority. This paper presents a fast method for global search of atomic structures at ab initio level. The structures global optimization (SGO) engine consists of a high-efficiency differential evolution algorithm, accelerated local relaxation methods and a plane-wave density functional theory code running on GPU machines. The purpose is to show what can be achieved by combining the superior algorithms at the different levels of the searching scheme. SGO can search the global-minimum configurations of crystals, two-dimensional materials and quantum clusters without prior symmetry restriction in a relatively short time (half or several hours for systems with less than 25 atoms), thus making such a task a routine calculation. Comparisons with other existing methods such as minima hopping and genetic algorithm are provided. One motivation of our study is to investigate the properties of magnetic systems in different phases. The SGO engine is capable of surveying the local minima surrounding the global minimum, which provides the information for the overall energy landscape of a given system. Using this capability we have found several new configurations for testing systems, explored their energy landscape, and demonstrated that the magnetic moment of metal clusters fluctuates strongly in different local minima.
Efficient algorithms for multidimensional global optimization in genetic mapping of complex traits
Directory of Open Access Journals (Sweden)
Kajsa Ljungberg
2010-10-01
Full Text Available Kajsa Ljungberg1, Kateryna Mishchenko2, Sverker Holmgren11Division of Scientific Computing, Department of Information Technology, Uppsala University, Uppsala, Sweden; 2Department of Mathematics and Physics, Mälardalen University College, Västerås, SwedenAbstract: We present a two-phase strategy for optimizing a multidimensional, nonconvex function arising during genetic mapping of quantitative traits. Such traits are believed to be affected by multiple so called QTL, and searching for d QTL results in a d-dimensional optimization problem with a large number of local optima. We combine the global algorithm DIRECT with a number of local optimization methods that accelerate the final convergence, and adapt the algorithms to problem-specific features. We also improve the evaluation of the QTL mapping objective function to enable exploitation of the smoothness properties of the optimization landscape. Our best two-phase method is demonstrated to be accurate in at least six dimensions and up to ten times faster than currently used QTL mapping algorithms.Keywords: global optimization, QTL mapping, DIRECT
International Nuclear Information System (INIS)
Frolov, A.M.
1986-01-01
The problem of exact variational calculations of few-particle systems in the exponential basis of the relative coordinates using nonlinear parameters is studied. The techniques of stepwise optimization and global chaos of nonlinear parameters are used to calculate the S and P states of homonuclear muonic molecules with an error of no more than +0.001 eV. The global-chaos technique also has proved to be successful in the case of the nuclear systems 3 H and 3 He
DEFF Research Database (Denmark)
Rasmussen, Marie-Louise Højlund; Stolpe, Mathias
2008-01-01
the physics, and the cuts (Combinatorial Benders’ and projected Chvátal–Gomory) come from an understanding of the particular mathematical structure of the reformulation. The impact of a stronger representation is investigated on several truss topology optimization problems in two and three dimensions.......The subject of this article is solving discrete truss topology optimization problems with local stress and displacement constraints to global optimum. We consider a formulation based on the Simultaneous ANalysis and Design (SAND) approach. This intrinsically non-convex problem is reformulated...
External costs in the global energy optimization models. A tool in favour of sustain ability
International Nuclear Information System (INIS)
Cabal Cuesta, H.
2007-01-01
The aim of this work is the analysis of the effects of the GHG external costs internalization in the energy systems. This may provide a useful tool to support decision makers to help reaching the energy systems sustain ability. External costs internalization has been carried out using two methods. First, CO 2 externalities of different power generation technologies have been internalized to evaluate their effects on the economic competitiveness of these present and future technologies. The other method consisted of analysing and optimizing the global energy system, from an economic and environmental point of view, using the global energy optimization model generator, TIMES, with a time horizon of 50 years. Finally, some scenarios regarding environmental and economic strategic measures have been analysed. (Author)
Memetic Algorithms to Solve a Global Nonlinear Optimization Problem. A Review
Directory of Open Access Journals (Sweden)
M. K. Sakharov
2015-01-01
Full Text Available In recent decades, evolutionary algorithms have proven themselves as the powerful optimization techniques of search engine. Their popularity is due to the fact that they are easy to implement and can be used in all areas, since they are based on the idea of universal evolution. For example, in the problems of a large number of local optima, the traditional optimization methods, usually, fail in finding the global optimum. To solve such problems using a variety of stochastic methods, in particular, the so-called population-based algorithms, which are a kind of evolutionary methods. The main disadvantage of this class of methods is their slow convergence to the exact solution in the neighborhood of the global optimum, as these methods incapable to use the local information about the landscape of the function. This often limits their use in largescale real-world problems where the computation time is a critical factor.One of the promising directions in the field of modern evolutionary computation are memetic algorithms, which can be regarded as a combination of population search of the global optimum and local procedures for verifying solutions, which gives a synergistic effect. In the context of memetic algorithms, the meme is an implementation of the local optimization method to refine solution in the search.The concept of memetic algorithms provides ample opportunities for the development of various modifications of these algorithms, which can vary the frequency of the local search, the conditions of its end, and so on. The practically significant memetic algorithm modifications involve the simultaneous use of different memes. Such algorithms are called multi-memetic.The paper gives statement of the global problem of nonlinear unconstrained optimization, describes the most promising areas of AI modifications, including hybridization and metaoptimization. The main content of the work is the classification and review of existing varieties of
Global optimization of proteins using a dynamical lattice model: Ground states and energy landscapes
Dressel, F.; Kobe, S.
2004-01-01
A simple approach is proposed to investigate the protein structure. Using a low complexity model, a simple pairwise interaction and the concept of global optimization, we are able to calculate ground states of proteins, which are in agreement with experimental data. All possible model structures of small proteins are available below a certain energy threshold. The exact lowenergy landscapes for the trp cage protein (1L2Y) is presented showing the connectivity of all states and energy barriers.
DEFF Research Database (Denmark)
Achtziger, Wolfgang; Stolpe, Mathias
2009-01-01
we use the theory developed in Part I to design a convergent nonlinear branch-and-bound method tailored to solve large-scale instances of the original discrete problem. The problem formulation and the needed theoretical results from Part I are repeated such that this paper is self-contained. We focus...... the largest discrete topology design problems solved by means of global optimization....
Zhong, Shangping; Chen, Tianshun; He, Fengying; Niu, Yuzhen
2014-09-01
For a practical pattern classification task solved by kernel methods, the computing time is mainly spent on kernel learning (or training). However, the current kernel learning approaches are based on local optimization techniques, and hard to have good time performances, especially for large datasets. Thus the existing algorithms cannot be easily extended to large-scale tasks. In this paper, we present a fast Gaussian kernel learning method by solving a specially structured global optimization (SSGO) problem. We optimize the Gaussian kernel function by using the formulated kernel target alignment criterion, which is a difference of increasing (d.i.) functions. Through using a power-transformation based convexification method, the objective criterion can be represented as a difference of convex (d.c.) functions with a fixed power-transformation parameter. And the objective programming problem can then be converted to a SSGO problem: globally minimizing a concave function over a convex set. The SSGO problem is classical and has good solvability. Thus, to find the global optimal solution efficiently, we can adopt the improved Hoffman's outer approximation method, which need not repeat the searching procedure with different starting points to locate the best local minimum. Also, the proposed method can be proven to converge to the global solution for any classification task. We evaluate the proposed method on twenty benchmark datasets, and compare it with four other Gaussian kernel learning methods. Experimental results show that the proposed method stably achieves both good time-efficiency performance and good classification performance. Copyright © 2014 Elsevier Ltd. All rights reserved.
Ringed Seal Search for Global Optimization via a Sensitive Search Model.
Directory of Open Access Journals (Sweden)
Younes Saadi
Full Text Available The efficiency of a metaheuristic algorithm for global optimization is based on its ability to search and find the global optimum. However, a good search often requires to be balanced between exploration and exploitation of the search space. In this paper, a new metaheuristic algorithm called Ringed Seal Search (RSS is introduced. It is inspired by the natural behavior of the seal pup. This algorithm mimics the seal pup movement behavior and its ability to search and choose the best lair to escape predators. The scenario starts once the seal mother gives birth to a new pup in a birthing lair that is constructed for this purpose. The seal pup strategy consists of searching and selecting the best lair by performing a random walk to find a new lair. Affected by the sensitive nature of seals against external noise emitted by predators, the random walk of the seal pup takes two different search states, normal state and urgent state. In the normal state, the pup performs an intensive search between closely adjacent lairs; this movement is modeled via a Brownian walk. In an urgent state, the pup leaves the proximity area and performs an extensive search to find a new lair from sparse targets; this movement is modeled via a Levy walk. The switch between these two states is realized by the random noise emitted by predators. The algorithm keeps switching between normal and urgent states until the global optimum is reached. Tests and validations were performed using fifteen benchmark test functions to compare the performance of RSS with other baseline algorithms. The results show that RSS is more efficient than Genetic Algorithm, Particles Swarm Optimization and Cuckoo Search in terms of convergence rate to the global optimum. The RSS shows an improvement in terms of balance between exploration (extensive and exploitation (intensive of the search space. The RSS can efficiently mimic seal pups behavior to find best lair and provide a new algorithm to be
Sequential Optimization of Global Sequence Alignments Relative to Different Cost Functions
Odat, Enas M.
2011-05-01
The purpose of this dissertation is to present a methodology to model global sequence alignment problem as directed acyclic graph which helps to extract all possible optimal alignments. Moreover, a mechanism to sequentially optimize sequence alignment problem relative to different cost functions is suggested. Sequence alignment is mostly important in computational biology. It is used to find evolutionary relationships between biological sequences. There are many algo- rithms that have been developed to solve this problem. The most famous algorithms are Needleman-Wunsch and Smith-Waterman that are based on dynamic program- ming. In dynamic programming, problem is divided into a set of overlapping sub- problems and then the solution of each subproblem is found. Finally, the solutions to these subproblems are combined into a final solution. In this thesis it has been proved that for two sequences of length m and n over a fixed alphabet, the suggested optimization procedure requires O(mn) arithmetic operations per cost function on a single processor machine. The algorithm has been simulated using C#.Net programming language and a number of experiments have been done to verify the proved statements. The results of these experiments show that the number of optimal alignments is reduced after each step of optimization. Furthermore, it has been verified that as the sequence length increased linearly then the number of optimal alignments increased exponentially which also depends on the cost function that is used. Finally, the number of executed operations increases polynomially as the sequence length increase linearly.
Economic optimization of a global strategy to address the pandemic threat.
Pike, Jamison; Bogich, Tiffany; Elwood, Sarah; Finnoff, David C; Daszak, Peter
2014-12-30
Emerging pandemics threaten global health and economies and are increasing in frequency. Globally coordinated strategies to combat pandemics, similar to current strategies that address climate change, are largely adaptive, in that they attempt to reduce the impact of a pathogen after it has emerged. However, like climate change, mitigation strategies have been developed that include programs to reduce the underlying drivers of pandemics, particularly animal-to-human disease transmission. Here, we use real options economic modeling of current globally coordinated adaptation strategies for pandemic prevention. We show that they would be optimally implemented within 27 y to reduce the annual rise of emerging infectious disease events by 50% at an estimated one-time cost of approximately $343.7 billion. We then analyze World Bank data on multilateral "One Health" pandemic mitigation programs. We find that, because most pandemics have animal origins, mitigation is a more cost-effective policy than business-as-usual adaptation programs, saving between $344.0.7 billion and $360.3 billion over the next 100 y if implemented today. We conclude that globally coordinated pandemic prevention policies need to be enacted urgently to be optimally effective and that strategies to mitigate pandemics by reducing the impact of their underlying drivers are likely to be more effective than business as usual.
Economic optimization of a global strategy to address the pandemic threat
Pike, Jamison; Bogich, Tiffany; Elwood, Sarah; Finnoff, David C.; Daszak, Peter
2014-01-01
Emerging pandemics threaten global health and economies and are increasing in frequency. Globally coordinated strategies to combat pandemics, similar to current strategies that address climate change, are largely adaptive, in that they attempt to reduce the impact of a pathogen after it has emerged. However, like climate change, mitigation strategies have been developed that include programs to reduce the underlying drivers of pandemics, particularly animal-to-human disease transmission. Here, we use real options economic modeling of current globally coordinated adaptation strategies for pandemic prevention. We show that they would be optimally implemented within 27 y to reduce the annual rise of emerging infectious disease events by 50% at an estimated one-time cost of approximately $343.7 billion. We then analyze World Bank data on multilateral “One Health” pandemic mitigation programs. We find that, because most pandemics have animal origins, mitigation is a more cost-effective policy than business-as-usual adaptation programs, saving between $344.0.7 billion and $360.3 billion over the next 100 y if implemented today. We conclude that globally coordinated pandemic prevention policies need to be enacted urgently to be optimally effective and that strategies to mitigate pandemics by reducing the impact of their underlying drivers are likely to be more effective than business as usual. PMID:25512538
SU-E-J-130: Automating Liver Segmentation Via Combined Global and Local Optimization
Energy Technology Data Exchange (ETDEWEB)
Li, Dengwang; Wang, Jie [College of Physics and Electronics, Shandong Normal University, Jinan, Shandong (China); Kapp, Daniel S.; Xing, Lei [Department of Radiation Oncology, Stanford University School of Medicine, Stanford, CA (United States)
2015-06-15
Purpose: The aim of this work is to develop a robust algorithm for accurate segmentation of liver with special attention paid to the problems with fuzzy edges and tumor. Methods: 200 CT images were collected from radiotherapy treatment planning system. 150 datasets are selected as the panel data for shape dictionary and parameters estimation. The remaining 50 datasets were used as test images. In our study liver segmentation was formulated as optimization process of implicit function. The liver region was optimized via local and global optimization during iterations. Our method consists five steps: 1)The livers from the panel data were segmented manually by physicians, and then We estimated the parameters of GMM (Gaussian mixture model) and MRF (Markov random field). Shape dictionary was built by utilizing the 3D liver shapes. 2)The outlines of chest and abdomen were located according to rib structure in the input images, and the liver region was initialized based on GMM. 3)The liver shape for each 2D slice was adjusted using MRF within the neighborhood of liver edge for local optimization. 4)The 3D liver shape was corrected by employing SSR (sparse shape representation) based on liver shape dictionary for global optimization. Furthermore, H-PSO(Hybrid Particle Swarm Optimization) was employed to solve the SSR equation. 5)The corrected 3D liver was divided into 2D slices as input data of the third step. The iteration was repeated within the local optimization and global optimization until it satisfied the suspension conditions (maximum iterations and changing rate). Results: The experiments indicated that our method performed well even for the CT images with fuzzy edge and tumors. Comparing with physician delineated results, the segmentation accuracy with the 50 test datasets (VOE, volume overlap percentage) was on average 91%–95%. Conclusion: The proposed automatic segmentation method provides a sensible technique for segmentation of CT images. This work is
SU-E-J-130: Automating Liver Segmentation Via Combined Global and Local Optimization
International Nuclear Information System (INIS)
Li, Dengwang; Wang, Jie; Kapp, Daniel S.; Xing, Lei
2015-01-01
Purpose: The aim of this work is to develop a robust algorithm for accurate segmentation of liver with special attention paid to the problems with fuzzy edges and tumor. Methods: 200 CT images were collected from radiotherapy treatment planning system. 150 datasets are selected as the panel data for shape dictionary and parameters estimation. The remaining 50 datasets were used as test images. In our study liver segmentation was formulated as optimization process of implicit function. The liver region was optimized via local and global optimization during iterations. Our method consists five steps: 1)The livers from the panel data were segmented manually by physicians, and then We estimated the parameters of GMM (Gaussian mixture model) and MRF (Markov random field). Shape dictionary was built by utilizing the 3D liver shapes. 2)The outlines of chest and abdomen were located according to rib structure in the input images, and the liver region was initialized based on GMM. 3)The liver shape for each 2D slice was adjusted using MRF within the neighborhood of liver edge for local optimization. 4)The 3D liver shape was corrected by employing SSR (sparse shape representation) based on liver shape dictionary for global optimization. Furthermore, H-PSO(Hybrid Particle Swarm Optimization) was employed to solve the SSR equation. 5)The corrected 3D liver was divided into 2D slices as input data of the third step. The iteration was repeated within the local optimization and global optimization until it satisfied the suspension conditions (maximum iterations and changing rate). Results: The experiments indicated that our method performed well even for the CT images with fuzzy edge and tumors. Comparing with physician delineated results, the segmentation accuracy with the 50 test datasets (VOE, volume overlap percentage) was on average 91%–95%. Conclusion: The proposed automatic segmentation method provides a sensible technique for segmentation of CT images. This work is
International Nuclear Information System (INIS)
Göktürkler, G; Balkaya, Ç
2012-01-01
Three naturally inspired meta-heuristic algorithms—the genetic algorithm (GA), simulated annealing (SA) and particle swarm optimization (PSO)—were used to invert some of the self-potential (SP) anomalies originated by some polarized bodies with simple geometries. Both synthetic and field data sets were considered. The tests with the synthetic data comprised of the solutions with both noise-free and noisy data; in the tests with the field data some SP anomalies observed over a copper belt (India), graphite deposits (Germany) and metallic sulfide (Turkey) were inverted. The model parameters included the electric dipole moment, polarization angle, depth, shape factor and origin of the anomaly. The estimated parameters were compared with those from previous studies using various optimization algorithms, mainly least-squares approaches, on the same data sets. During the test studies the solutions by GA, PSO and SA were characterized as being consistent with each other; a good starting model was not a requirement to reach the global minimum. It can be concluded that the global optimization algorithms considered in this study were able to yield compatible solutions with those from widely used local optimization algorithms. (paper)
Annealing evolutionary stochastic approximation Monte Carlo for global optimization
Liang, Faming
2010-04-08
In this paper, we propose a new algorithm, the so-called annealing evolutionary stochastic approximation Monte Carlo (AESAMC) algorithm as a general optimization technique, and study its convergence. AESAMC possesses a self-adjusting mechanism, whose target distribution can be adapted at each iteration according to the current samples. Thus, AESAMC falls into the class of adaptive Monte Carlo methods. This mechanism also makes AESAMC less trapped by local energy minima than nonadaptive MCMC algorithms. Under mild conditions, we show that AESAMC can converge weakly toward a neighboring set of global minima in the space of energy. AESAMC is tested on multiple optimization problems. The numerical results indicate that AESAMC can potentially outperform simulated annealing, the genetic algorithm, annealing stochastic approximation Monte Carlo, and some other metaheuristics in function optimization. © 2010 Springer Science+Business Media, LLC.
Global optimization methods for the aerodynamic shape design of transonic cascades
International Nuclear Information System (INIS)
Mengistu, T.; Ghaly, W.
2003-01-01
Two global optimization algorithms, namely Genetic Algorithm (GA) and Simulated Annealing (SA), have been applied to the aerodynamic shape optimization of transonic cascades; the objective being the redesign of an existing turbomachine airfoil to improve its performance by minimizing the total pressure loss while satisfying a number of constraints. This is accomplished by modifying the blade camber line; keeping the same blade thickness distribution, mass flow rate and the same flow turning. The objective is calculated based on an Euler solver and the blade camber line is represented with non-uniform rational B-splines (NURBS). The SA and GA methods were first assessed for known test functions and their performance in optimizing the blade shape for minimum loss is then demonstrated on a transonic turbine cascade where it is shown to produce a significant reduction in total pressure loss by eliminating the passage shock. (author)
Institute of Scientific and Technical Information of China (English)
葛恒武; 陈中文
2002-01-01
We present a class of nonmonotone trust region algorithms for linearly constrained optimization in this paper.The algorithm may adjust automatically the scope of the monotonicity by the degree that the quadratic model is "trusted".Under the suitable conditions,it is proved that any limit point of the infinite sequence generated by the algorithm is the Kuhn-Tucker point of the primal problem.Finally,some numerical results show that the new algorithm is very effective.
A global carbon assimilation system based on a dual optimization method
Zheng, H.; Li, Y.; Chen, J. M.; Wang, T.; Huang, Q.; Huang, W. X.; Wang, L. H.; Li, S. M.; Yuan, W. P.; Zheng, X.; Zhang, S. P.; Chen, Z. Q.; Jiang, F.
2015-02-01
Ecological models are effective tools for simulating the distribution of global carbon sources and sinks. However, these models often suffer from substantial biases due to inaccurate simulations of complex ecological processes. We introduce a set of scaling factors (parameters) to an ecological model on the basis of plant functional type (PFT) and latitudes. A global carbon assimilation system (GCAS-DOM) is developed by employing a dual optimization method (DOM) to invert the time-dependent ecological model parameter state and the net carbon flux state simultaneously. We use GCAS-DOM to estimate the global distribution of the CO2 flux on 1° × 1° grid cells for the period from 2001 to 2007. Results show that land and ocean absorb -3.63 ± 0.50 and -1.82 ± 0.16 Pg C yr-1, respectively. North America, Europe and China contribute -0.98 ± 0.15, -0.42 ± 0.08 and -0.20 ± 0.29 Pg C yr-1, respectively. The uncertainties in the flux after optimization by GCAS-DOM have been remarkably reduced by more than 60%. Through parameter optimization, GCAS-DOM can provide improved estimates of the carbon flux for each PFT. Coniferous forest (-0.97 ± 0.27 Pg C yr-1) is the largest contributor to the global carbon sink. Fluxes of once-dominant deciduous forest generated by the Boreal Ecosystems Productivity Simulator (BEPS) are reduced to -0.78 ± 0.23 Pg C yr-1, the third largest carbon sink.
DEFF Research Database (Denmark)
Stolpe, Mathias; Bendsøe, Martin P.
2007-01-01
This paper present some initial results pertaining to a search for globally optimal solutions to a challenging benchmark example proposed by Zhou and Rozvany. This means that we are dealing with global optimization of the classical single load minimum compliance topology design problem with a fixed...... finite element discretization and with discrete design variables. Global optimality is achieved by the implementation of some specially constructed convergent nonlinear branch and cut methods, based on the use of natural relaxations and by applying strengthening constraints (linear valid inequalities...
DEFF Research Database (Denmark)
Stolpe, Mathias; Bendsøe, Martin P.
2007-01-01
This paper present some initial results pertaining to a search for globally optimal solutions to a challenging benchmark example proposed by Zhou and Rozvany. This means that we are dealing with global optimization of the classical single load minimum compliance topology design problem with a fixed...... finite element discretization and with discrete design variables. Global optimality is achieved by the implementation of some specially constructed convergent nonlinear branch and cut methods, based on the use of natural relaxations and by applying strengthening constraints (linear valid inequalities......) and cuts....
Global-Local Analysis and Optimization of a Composite Civil Tilt-Rotor Wing
Rais-Rohani, Masound
1999-01-01
This report gives highlights of an investigation on the design and optimization of a thin composite wing box structure for a civil tilt-rotor aircraft. Two different concepts are considered for the cantilever wing: (a) a thin monolithic skin design, and (b) a thick sandwich skin design. Each concept is examined with three different skin ply patterns based on various combinations of 0, +/-45, and 90 degree plies. The global-local technique is used in the analysis and optimization of the six design models. The global analysis is based on a finite element model of the wing-pylon configuration while the local analysis uses a uniformly supported plate representing a wing panel. Design allowables include those on vibration frequencies, panel buckling, and material strength. The design optimization problem is formulated as one of minimizing the structural weight subject to strength, stiffness, and d,vnamic constraints. Six different loading conditions based on three different flight modes are considered in the design optimization. The results of this investigation reveal that of all the loading conditions the one corresponding to the rolling pull-out in the airplane mode is the most stringent. Also the frequency constraints are found to drive the skin thickness limits, rendering the buckling constraints inactive. The optimum skin ply pattern for the monolithic skin concept is found to be (((0/+/-45/90/(0/90)(sub 2))(sub s))(sub s), while for the sandwich skin concept the optimal ply pattern is found to be ((0/+/-45/90)(sub 2s))(sub s).
Optimizing rice yields while minimizing yield-scaled global warming potential.
Pittelkow, Cameron M; Adviento-Borbe, Maria A; van Kessel, Chris; Hill, James E; Linquist, Bruce A
2014-05-01
To meet growing global food demand with limited land and reduced environmental impact, agricultural greenhouse gas (GHG) emissions are increasingly evaluated with respect to crop productivity, i.e., on a yield-scaled as opposed to area basis. Here, we compiled available field data on CH4 and N2 O emissions from rice production systems to test the hypothesis that in response to fertilizer nitrogen (N) addition, yield-scaled global warming potential (GWP) will be minimized at N rates that maximize yields. Within each study, yield N surplus was calculated to estimate deficit or excess N application rates with respect to the optimal N rate (defined as the N rate at which maximum yield was achieved). Relationships between yield N surplus and GHG emissions were assessed using linear and nonlinear mixed-effects models. Results indicate that yields increased in response to increasing N surplus when moving from deficit to optimal N rates. At N rates contributing to a yield N surplus, N2 O and yield-scaled N2 O emissions increased exponentially. In contrast, CH4 emissions were not impacted by N inputs. Accordingly, yield-scaled CH4 emissions decreased with N addition. Overall, yield-scaled GWP was minimized at optimal N rates, decreasing by 21% compared to treatments without N addition. These results are unique compared to aerobic cropping systems in which N2 O emissions are the primary contributor to GWP, meaning yield-scaled GWP may not necessarily decrease for aerobic crops when yields are optimized by N fertilizer addition. Balancing gains in agricultural productivity with climate change concerns, this work supports the concept that high rice yields can be achieved with minimal yield-scaled GWP through optimal N application rates. Moreover, additional improvements in N use efficiency may further reduce yield-scaled GWP, thereby strengthening the economic and environmental sustainability of rice systems. © 2013 John Wiley & Sons Ltd.
Comparison of global optimization approaches for robust calibration of hydrologic model parameters
Jung, I. W.
2015-12-01
Robustness of the calibrated parameters of hydrologic models is necessary to provide a reliable prediction of future performance of watershed behavior under varying climate conditions. This study investigated calibration performances according to the length of calibration period, objective functions, hydrologic model structures and optimization methods. To do this, the combination of three global optimization methods (i.e. SCE-UA, Micro-GA, and DREAM) and four hydrologic models (i.e. SAC-SMA, GR4J, HBV, and PRMS) was tested with different calibration periods and objective functions. Our results showed that three global optimization methods provided close calibration performances under different calibration periods, objective functions, and hydrologic models. However, using the agreement of index, normalized root mean square error, Nash-Sutcliffe efficiency as the objective function showed better performance than using correlation coefficient and percent bias. Calibration performances according to different calibration periods from one year to seven years were hard to generalize because four hydrologic models have different levels of complexity and different years have different information content of hydrological observation. Acknowledgements This research was supported by a grant (14AWMP-B082564-01) from Advanced Water Management Research Program funded by Ministry of Land, Infrastructure and Transport of Korean government.
QuickVina: accelerating AutoDock Vina using gradient-based heuristics for global optimization.
Handoko, Stephanus Daniel; Ouyang, Xuchang; Su, Chinh Tran To; Kwoh, Chee Keong; Ong, Yew Soon
2012-01-01
Predicting binding between macromolecule and small molecule is a crucial phase in the field of rational drug design. AutoDock Vina, one of the most widely used docking software released in 2009, uses an empirical scoring function to evaluate the binding affinity between the molecules and employs the iterated local search global optimizer for global optimization, achieving a significantly improved speed and better accuracy of the binding mode prediction compared its predecessor, AutoDock 4. In this paper, we propose further improvement in the local search algorithm of Vina by heuristically preventing some intermediate points from undergoing local search. Our improved version of Vina-dubbed QVina-achieved a maximum acceleration of about 25 times with the average speed-up of 8.34 times compared to the original Vina when tested on a set of 231 protein-ligand complexes while maintaining the optimal scores mostly identical. Using our heuristics, larger number of different ligands can be quickly screened against a given receptor within the same time frame.
Liang, Faming
2014-04-03
Simulated annealing has been widely used in the solution of optimization problems. As known by many researchers, the global optima cannot be guaranteed to be located by simulated annealing unless a logarithmic cooling schedule is used. However, the logarithmic cooling schedule is so slow that no one can afford to use this much CPU time. This article proposes a new stochastic optimization algorithm, the so-called simulated stochastic approximation annealing algorithm, which is a combination of simulated annealing and the stochastic approximation Monte Carlo algorithm. Under the framework of stochastic approximation, it is shown that the new algorithm can work with a cooling schedule in which the temperature can decrease much faster than in the logarithmic cooling schedule, for example, a square-root cooling schedule, while guaranteeing the global optima to be reached when the temperature tends to zero. The new algorithm has been tested on a few benchmark optimization problems, including feed-forward neural network training and protein-folding. The numerical results indicate that the new algorithm can significantly outperform simulated annealing and other competitors. Supplementary materials for this article are available online.
Directory of Open Access Journals (Sweden)
JongHyup Lee
2016-08-01
Full Text Available For practical deployment of wireless sensor networks (WSN, WSNs construct clusters, where a sensor node communicates with other nodes in its cluster, and a cluster head support connectivity between the sensor nodes and a sink node. In hybrid WSNs, cluster heads have cellular network interfaces for global connectivity. However, when WSNs are active and the load of cellular networks is high, the optimal assignment of cluster heads to base stations becomes critical. Therefore, in this paper, we propose a game theoretic model to find the optimal assignment of base stations for hybrid WSNs. Since the communication and energy cost is different according to cellular systems, we devise two game models for TDMA/FDMA and CDMA systems employing power prices to adapt to the varying efficiency of recent wireless technologies. The proposed model is defined on the assumptions of the ideal sensing field, but our evaluation shows that the proposed model is more adaptive and energy efficient than local selections.
Lee, JongHyup; Pak, Dohyun
2016-01-01
For practical deployment of wireless sensor networks (WSN), WSNs construct clusters, where a sensor node communicates with other nodes in its cluster, and a cluster head support connectivity between the sensor nodes and a sink node. In hybrid WSNs, cluster heads have cellular network interfaces for global connectivity. However, when WSNs are active and the load of cellular networks is high, the optimal assignment of cluster heads to base stations becomes critical. Therefore, in this paper, we propose a game theoretic model to find the optimal assignment of base stations for hybrid WSNs. Since the communication and energy cost is different according to cellular systems, we devise two game models for TDMA/FDMA and CDMA systems employing power prices to adapt to the varying efficiency of recent wireless technologies. The proposed model is defined on the assumptions of the ideal sensing field, but our evaluation shows that the proposed model is more adaptive and energy efficient than local selections. PMID:27589743
Research on optimal investment path of transmission corridor under the global energy Internet
Huang, Yuehui; Li, Pai; Wang, Qi; Liu, Jichun; Gao, Han
2018-02-01
Under the background of the global energy Internet, the investment planning of transmission corridor from XinJiang to Germany is studied in this article, which passes through four countries: Kazakhstan, Russia, Belarus and Poland. Taking the specific situation of different countries into account, including the length of transmission line, unit construction cost, completion time, transmission price, state tariff, inflation rate and so on, this paper constructed a power transmission investment model. Finally, the dynamic programming method is used to simulate the example, and the optimal strategies under different objective functions are obtained.
Global stability, periodic solutions, and optimal control in a nonlinear differential delay model
Directory of Open Access Journals (Sweden)
Anatoli F. Ivanov
2010-09-01
Full Text Available A nonlinear differential equation with delay serving as a mathematical model of several applied problems is considered. Sufficient conditions for the global asymptotic stability and for the existence of periodic solutions are given. Two particular applications are treated in detail. The first one is a blood cell production model by Mackey, for which new periodicity criteria are derived. The second application is a modified economic model with delay due to Ramsey. An optimization problem for a maximal consumption is stated and solved for the latter.
Global Convergence of a Spectral Conjugate Gradient Method for Unconstrained Optimization
Directory of Open Access Journals (Sweden)
Jinkui Liu
2012-01-01
Full Text Available A new nonlinear spectral conjugate descent method for solving unconstrained optimization problems is proposed on the basis of the CD method and the spectral conjugate gradient method. For any line search, the new method satisfies the sufficient descent condition gkTdk<−∥gk∥2. Moreover, we prove that the new method is globally convergent under the strong Wolfe line search. The numerical results show that the new method is more effective for the given test problems from the CUTE test problem library (Bongartz et al., 1995 in contrast to the famous CD method, FR method, and PRP method.
Miró, Anton; Pozo, Carlos; Guillén-Gosálbez, Gonzalo; Egea, Jose A; Jiménez, Laureano
2012-05-10
The estimation of parameter values for mathematical models of biological systems is an optimization problem that is particularly challenging due to the nonlinearities involved. One major difficulty is the existence of multiple minima in which standard optimization methods may fall during the search. Deterministic global optimization methods overcome this limitation, ensuring convergence to the global optimum within a desired tolerance. Global optimization techniques are usually classified into stochastic and deterministic. The former typically lead to lower CPU times but offer no guarantee of convergence to the global minimum in a finite number of iterations. In contrast, deterministic methods provide solutions of a given quality (i.e., optimality gap), but tend to lead to large computational burdens. This work presents a deterministic outer approximation-based algorithm for the global optimization of dynamic problems arising in the parameter estimation of models of biological systems. Our approach, which offers a theoretical guarantee of convergence to global minimum, is based on reformulating the set of ordinary differential equations into an equivalent set of algebraic equations through the use of orthogonal collocation methods, giving rise to a nonconvex nonlinear programming (NLP) problem. This nonconvex NLP is decomposed into two hierarchical levels: a master mixed-integer linear programming problem (MILP) that provides a rigorous lower bound on the optimal solution, and a reduced-space slave NLP that yields an upper bound. The algorithm iterates between these two levels until a termination criterion is satisfied. The capabilities of our approach were tested in two benchmark problems, in which the performance of our algorithm was compared with that of the commercial global optimization package BARON. The proposed strategy produced near optimal solutions (i.e., within a desired tolerance) in a fraction of the CPU time required by BARON.
Energy Technology Data Exchange (ETDEWEB)
T. Hikmet Karakoc; Onder Turan [School of Civil Aviation, Anadolu University, Eskisehir (Turkey)
2008-09-30
The main objective of the present study is to perform minimizing specific fuel consumption of a non afterburning high bypass turbofan engine with separate exhaust streams and unmixed flow for reducing global effect. The values of engine design parameters are optimized for maintaining minimum specific fuel consumption of high bypass turbofan engine under different flight conditions, different fuel types and design criteria. The backbones of optimization approach consisted of elitism-based genetic algorithm coupled with real parametric cycle analysis of a turbofan engine. For solving optimization problem a new software program is developed in MATLAB programming language, while objective function is determined for minimizing the specific fuel consumption. The input variables included the compressor pressure ratio ({pi}{sub c}), bypass ratio ({alpha}) and the fuel heating value [h{sub PR}-(kJ/kg)]. Hydrogen was selected as fuel type in real parametric cycle analysis of commercial turbofans. It may be concluded that the software program developed can successfully solve optimization problems at 10{le}{pi}{sub c}{le}20, 2{le}{alpha}{le}10 and h{sub PR} 120,000 with aircraft flight Mach number {le}0.8.
Guiding automated NMR structure determination using a global optimization metric, the NMR DP score
Energy Technology Data Exchange (ETDEWEB)
Huang, Yuanpeng Janet, E-mail: yphuang@cabm.rutgers.edu; Mao, Binchen; Xu, Fei; Montelione, Gaetano T., E-mail: gtm@rutgers.edu [Rutgers, The State University of New Jersey, Department of Molecular Biology and Biochemistry, Center for Advanced Biotechnology and Medicine, and Northeast Structural Genomics Consortium (United States)
2015-08-15
ASDP is an automated NMR NOE assignment program. It uses a distinct bottom-up topology-constrained network anchoring approach for NOE interpretation, with 2D, 3D and/or 4D NOESY peak lists and resonance assignments as input, and generates unambiguous NOE constraints for iterative structure calculations. ASDP is designed to function interactively with various structure determination programs that use distance restraints to generate molecular models. In the CASD–NMR project, ASDP was tested and further developed using blinded NMR data, including resonance assignments, either raw or manually-curated (refined) NOESY peak list data, and in some cases {sup 15}N–{sup 1}H residual dipolar coupling data. In these blinded tests, in which the reference structure was not available until after structures were generated, the fully-automated ASDP program performed very well on all targets using both the raw and refined NOESY peak list data. Improvements of ASDP relative to its predecessor program for automated NOESY peak assignments, AutoStructure, were driven by challenges provided by these CASD–NMR data. These algorithmic improvements include (1) using a global metric of structural accuracy, the discriminating power score, for guiding model selection during the iterative NOE interpretation process, and (2) identifying incorrect NOESY cross peak assignments caused by errors in the NMR resonance assignment list. These improvements provide a more robust automated NOESY analysis program, ASDP, with the unique capability of being utilized with alternative structure generation and refinement programs including CYANA, CNS, and/or Rosetta.
Energy Technology Data Exchange (ETDEWEB)
Fiorucci, I.; Muscari, G. [Istituto Nazionale di Geofisica e Vulcanologia, Rome (Italy); De Zafra, R.L. [State Univ. of New York, Stony Brook, NY (United States). Dept. of Physics and Astronomy
2011-07-01
The Ground-Based Millimeter-wave Spectrometer (GBMS) was designed and built at the State University of New York at Stony Brook in the early 1990s and since then has carried out many measurement campaigns of stratospheric O{sub 3}, HNO{sub 3}, CO and N{sub 2}O at polar and mid-latitudes. Its HNO{sub 3} data set shed light on HNO{sub 3} annual cycles over the Antarctic continent and contributed to the validation of both generations of the satellite-based JPL Microwave Limb Sounder (MLS). Following the increasing need for long-term data sets of stratospheric constituents, we resolved to establish a long-term GMBS observation site at the Arctic station of Thule (76.5 N, 68.8 W), Greenland, beginning in January 2009, in order to track the long- and short-term interactions between the changing climate and the seasonal processes tied to the ozone depletion phenomenon. Furthermore, we updated the retrieval algorithm adapting the Optimal Estimation (OE) method to GBMS spectral data in order to conform to the standard of the Network for the Detection of Atmospheric Composition Change (NDACC) microwave group, and to provide our retrievals with a set of averaging kernels that allow more straightforward comparisons with other data sets. The new OE algorithm was applied to GBMS HNO{sub 3} data sets from 1993 South Pole observations to date, in order to produce HNO{sub 3} version 2 (v2) profiles. A sample of results obtained at Antarctic latitudes in fall and winter and at mid-latitudes is shown here. In most conditions, v2 inversions show a sensitivity (i.e., sum of column elements of the averaging kernel matrix) of 100{+-}20% from 20 to 45 km altitude, with somewhat worse (better) sensitivity in the Antarctic winter lower (upper) stratosphere. The 1{sigma} uncertainty on HNO{sub 3} v2 mixing ratio vertical profiles depends on altitude and is estimated at {proportional_to}15% or 0.3 ppbv, whichever is larger. Comparisons of v2 with former (v1) GBMS HNO{sub 3} vertical profiles
The Multipoint Global Shape Optimization of Flying Configuration with Movable Leading Edges Flaps
Directory of Open Access Journals (Sweden)
Adriana NASTASE
2012-12-01
Full Text Available The aerodynamical global optimized (GO shape of flying configuration (FC, at two cruising Mach numbers, can be realized by morphing. Movable leading edge flaps are used for this purpose. The equations of the surfaces of the wing, of the fuselage and of the flaps in stretched position are approximated in form of superpositions of homogeneous polynomes in two variables with free coefficients. These coefficients together with the similarity parameters of the planform of the FC are the free parameters of the global optimization. Two enlarged variational problems with free boundaries occur. The first one consists in the determination of the GO shape of the wing-fuselageFC, with the flaps in retracted position, which must be of minimum drag, at higher cruising Mach number. The second enlarged variational problem consists in the determination of the GO shape of the flaps in stretched position in such a manner that the entire FC shall be of minimum drag at the second lower Mach number. The iterative optimum-optimorum (OO theory of the author is used for the solving of these both enlarged variational problems. The inviscid GO shape of the FC is used only in the first step of iteration and the own developed hybrid solutions for the compressible Navier-Stokes partial-differential equations (PDEs are used for the determination of the friction drag coefficient and up the second step of iteration of OO theory.
An efficient global energy optimization approach for robust 3D plane segmentation of point clouds
Dong, Zhen; Yang, Bisheng; Hu, Pingbo; Scherer, Sebastian
2018-03-01
Automatic 3D plane segmentation is necessary for many applications including point cloud registration, building information model (BIM) reconstruction, simultaneous localization and mapping (SLAM), and point cloud compression. However, most of the existing 3D plane segmentation methods still suffer from low precision and recall, and inaccurate and incomplete boundaries, especially for low-quality point clouds collected by RGB-D sensors. To overcome these challenges, this paper formulates the plane segmentation problem as a global energy optimization because it is robust to high levels of noise and clutter. First, the proposed method divides the raw point cloud into multiscale supervoxels, and considers planar supervoxels and individual points corresponding to nonplanar supervoxels as basic units. Then, an efficient hybrid region growing algorithm is utilized to generate initial plane set by incrementally merging adjacent basic units with similar features. Next, the initial plane set is further enriched and refined in a mutually reinforcing manner under the framework of global energy optimization. Finally, the performances of the proposed method are evaluated with respect to six metrics (i.e., plane precision, plane recall, under-segmentation rate, over-segmentation rate, boundary precision, and boundary recall) on two benchmark datasets. Comprehensive experiments demonstrate that the proposed method obtained good performances both in high-quality TLS point clouds (i.e., http://SEMANTIC3D.NET)
Protein structure modeling for CASP10 by multiple layers of global optimization.
Joo, Keehyoung; Lee, Juyong; Sim, Sangjin; Lee, Sun Young; Lee, Kiho; Heo, Seungryong; Lee, In-Ho; Lee, Sung Jong; Lee, Jooyoung
2014-02-01
In the template-based modeling (TBM) category of CASP10 experiment, we introduced a new protocol called protein modeling system (PMS) to generate accurate protein structures in terms of side-chains as well as backbone trace. In the new protocol, a global optimization algorithm, called conformational space annealing (CSA), is applied to the three layers of TBM procedure: multiple sequence-structure alignment, 3D chain building, and side-chain re-modeling. For 3D chain building, we developed a new energy function which includes new distance restraint terms of Lorentzian type (derived from multiple templates), and new energy terms that combine (physical) energy terms such as dynamic fragment assembly (DFA) energy, DFIRE statistical potential energy, hydrogen bonding term, etc. These physical energy terms are expected to guide the structure modeling especially for loop regions where no template structures are available. In addition, we developed a new quality assessment method based on random forest machine learning algorithm to screen templates, multiple alignments, and final models. For TBM targets of CASP10, we find that, due to the combination of three stages of CSA global optimizations and quality assessment, the modeling accuracy of PMS improves at each additional stage of the protocol. It is especially noteworthy that the side-chains of the final PMS models are far more accurate than the models in the intermediate steps. Copyright © 2013 Wiley Periodicals, Inc.
Directory of Open Access Journals (Sweden)
Abdulbaset El Hadi Saad
2017-10-01
Full Text Available Advanced global optimization algorithms have been continuously introduced and improved to solve various complex design optimization problems for which the objective and constraint functions can only be evaluated through computation intensive numerical analyses or simulations with a large number of design variables. The often implicit, multimodal, and ill-shaped objective and constraint functions in high-dimensional and “black-box” forms demand the search to be carried out using low number of function evaluations with high search efficiency and good robustness. This work investigates the performance of six recently introduced, nature-inspired global optimization methods: Artificial Bee Colony (ABC, Firefly Algorithm (FFA, Cuckoo Search (CS, Bat Algorithm (BA, Flower Pollination Algorithm (FPA and Grey Wolf Optimizer (GWO. These approaches are compared in terms of search efficiency and robustness in solving a set of representative benchmark problems in smooth-unimodal, non-smooth unimodal, smooth multimodal, and non-smooth multimodal function forms. In addition, four classic engineering optimization examples and a real-life complex mechanical system design optimization problem, floating offshore wind turbines design optimization, are used as additional test cases representing computationally-expensive black-box global optimization problems. Results from this comparative study show that the ability of these global optimization methods to obtain a good solution diminishes as the dimension of the problem, or number of design variables increases. Although none of these methods is universally capable, the study finds that GWO and ABC are more efficient on average than the other four in obtaining high quality solutions efficiently and consistently, solving 86% and 80% of the tested benchmark problems, respectively. The research contributes to future improvements of global optimization methods.
International Nuclear Information System (INIS)
Jiang, He; Dong, Yao; Wang, Jianzhou; Li, Yuqin
2015-01-01
Highlights: • CS-hard-ridge-RBF and DE-hard-ridge-RBF are proposed to forecast solar radiation. • Pearson and Apriori algorithm are used to analyze correlations between the data. • Hard-ridge penalty is added to reduce the number of nodes in the hidden layer. • CS algorithm and DE algorithm are used to determine the optimal parameters. • Proposed two models have higher forecasting accuracy than RBF and hard-ridge-RBF. - Abstract: Due to the scarcity of equipment and the high costs of maintenance, far fewer observations of solar radiation are made than observations of temperature, precipitation and other weather factors. Therefore, it is increasingly important to study several relevant meteorological factors to accurately forecast solar radiation. For this research, monthly average global solar radiation and 12 meteorological parameters from 1998 to 2010 at four sites in the United States were collected. Pearson correlation coefficients and Apriori association rules were successfully used to analyze correlations between the data, which provided a basis for these relative parameters as input variables. Two effective and innovative methods were developed to forecast monthly average global solar radiation by converting a RBF neural network into a multiple linear regression problem, adding a hard-ridge penalty to reduce the number of nodes in the hidden layer, and applying intelligent optimization algorithms, such as the cuckoo search algorithm (CS) and differential evolution (DE), to determine the optimal center and scale parameters. The experimental results show that the proposed models produce much more accurate forecasts than other models
Negotiation and Optimality in an Economic Model of Global Climate Change
Energy Technology Data Exchange (ETDEWEB)
Gottinger, H. [International Institute for Environmental Economics and Management IIEEM, University of Maastricht, Maastricht (Netherlands)
2000-03-01
The paper addresses the problem of governmental intervention in a multi-country regime of controlling global climate change. Using a simplified case of a two-country, two-sector general equilibrium model the paper shows that the global optimal time path of economic outputs and temperature will converge to a unique steady state provided that consumers care enough about the future. To answer a set of questions relating to 'what will happen if governments decide to correct the problem of global warming?' we study the equilibrium outcome in a bargaining game where two countries negotiate an agreement on future consumption and production plans for the purpose of correcting the problem of climate change. It is shown that the agreement arising from such a negotiation process achieves the best outcome and that it can be implemented in decentralised economies by a system of taxes, subsidies and transfers. By employing the recent advances in non-cooperative bargaining theory, the agreement between two countries is derived endogenously through a well-specified bargaining procedure.
Negotiation and Optimality in an Economic Model of Global Climate Change
International Nuclear Information System (INIS)
Gottinger, H.
2000-03-01
The paper addresses the problem of governmental intervention in a multi-country regime of controlling global climate change. Using a simplified case of a two-country, two-sector general equilibrium model the paper shows that the global optimal time path of economic outputs and temperature will converge to a unique steady state provided that consumers care enough about the future. To answer a set of questions relating to 'what will happen if governments decide to correct the problem of global warming?' we study the equilibrium outcome in a bargaining game where two countries negotiate an agreement on future consumption and production plans for the purpose of correcting the problem of climate change. It is shown that the agreement arising from such a negotiation process achieves the best outcome and that it can be implemented in decentralised economies by a system of taxes, subsidies and transfers. By employing the recent advances in non-cooperative bargaining theory, the agreement between two countries is derived endogenously through a well-specified bargaining procedure
Prediction of energy demands using neural network with model identification by global optimization
Energy Technology Data Exchange (ETDEWEB)
Yokoyama, Ryohei; Wakui, Tetsuya; Satake, Ryoichi [Department of Mechanical Engineering, Osaka Prefecture University, 1-1 Gakuen-cho, Naka-ku, Sakai, Osaka 599-8531 (Japan)
2009-02-15
To operate energy supply plants properly from the viewpoints of stable energy supply, and energy and cost savings, it is important to predict energy demands accurately as basic conditions. Several methods of predicting energy demands have been proposed, and one of them is to use neural networks. Although local optimization methods such as gradient ones have conventionally been adopted in the back propagation procedure to identify the values of model parameters, they have the significant drawback that they can derive only local optimal solutions. In this paper, a global optimization method called ''Modal Trimming Method'' proposed for non-linear programming problems is adopted to identify the values of model parameters. In addition, the trend and periodic change are first removed from time series data on energy demand, and the converted data is used as the main input to a neural network. Furthermore, predicted values of air temperature and relative humidity are considered as additional inputs to the neural network, and their effect on the prediction of energy demand is investigated. This approach is applied to the prediction of the cooling demand in a building used for a bench mark test of a variety of prediction methods, and its validity and effectiveness are clarified. (author)
Murillo, Sergio; Pattichis, Marios; Soliz, Peter; Barriga, Simon; Loizou, C. P.; Pattichis, C. S.
2010-03-01
Motion estimation from digital video is an ill-posed problem that requires a regularization approach. Regularization introduces a smoothness constraint that can reduce the resolution of the velocity estimates. The problem is further complicated for ultrasound videos (US), where speckle noise levels can be significant. Motion estimation using optical flow models requires the modification of several parameters to satisfy the optical flow constraint as well as the level of imposed smoothness. Furthermore, except in simulations or mostly unrealistic cases, there is no ground truth to use for validating the velocity estimates. This problem is present in all real video sequences that are used as input to motion estimation algorithms. It is also an open problem in biomedical applications like motion analysis of US of carotid artery (CA) plaques. In this paper, we study the problem of obtaining reliable ultrasound video motion estimates for atherosclerotic plaques for use in clinical diagnosis. A global optimization framework for motion parameter optimization is presented. This framework uses actual carotid artery motions to provide optimal parameter values for a variety of motions and is tested on ten different US videos using two different motion estimation techniques.
Lagos, Soledad R.; Velis, Danilo R.
2018-02-01
We perform the location of microseismic events generated in hydraulic fracturing monitoring scenarios using two global optimization techniques: Very Fast Simulated Annealing (VFSA) and Particle Swarm Optimization (PSO), and compare them against the classical grid search (GS). To this end, we present an integrated and optimized workflow that concatenates into an automated bash script the different steps that lead to the microseismic events location from raw 3C data. First, we carry out the automatic detection, denoising and identification of the P- and S-waves. Secondly, we estimate their corresponding backazimuths using polarization information, and propose a simple energy-based criterion to automatically decide which is the most reliable estimate. Finally, after taking proper care of the size of the search space using the backazimuth information, we perform the location using the aforementioned algorithms for 2D and 3D usual scenarios of hydraulic fracturing processes. We assess the impact of restricting the search space and show the advantages of using either VFSA or PSO over GS to attain significant speed-ups.
Libraro, Paola
The general electric propulsion orbit-raising maneuver of a spacecraft must contend with four main limiting factors: the longer time of flight, multiple eclipses prohibiting continuous thrusting, long exposure to radiation from the Van Allen belt and high power requirement of the electric engines. In order to optimize a low-thrust transfer with respect to these challenges, the choice of coordinates and corresponding equations of motion used to describe the kinematical and dynamical behavior of the satellite is of critical importance. This choice can potentially affect the numerical optimization process as well as limit the set of mission scenarios that can be investigated. To increase the ability to determine the feasible set of mission scenarios able to address the challenges of an all-electric orbit-raising, a set of equations free of any singularities is required to consider a completely arbitrary injection orbit. For this purpose a new quaternion-based formulation of a spacecraft translational dynamics that is globally nonsingular has been developed. The minimum-time low-thrust problem has been solved using the new set of equations of motion inside a direct optimization scheme in order to investigate optimal low-thrust trajectories over the full range of injection orbit inclinations between 0 and 90 degrees with particular focus on high-inclinations. The numerical results consider a specific mission scenario in order to analyze three key aspects of the problem: the effect of the initial guess on the shape and duration of the transfer, the effect of Earth oblateness on transfer time and the role played by, radiation damage and power degradation in all-electric minimum-time transfers. Finally trade-offs between mass and cost savings are introduced through a test case.
Population Structures in Russia: Optimality and Dependence on Parameters of Global Evolution
Directory of Open Access Journals (Sweden)
Yuri Yegorov
2016-07-01
Full Text Available The paper is devoted to analytical investigation of the division of geographical space into urban and rural areas with application to Russia. Yegorov (2005, 2006, 2009 has suggested the role of population density on economics. A city has an attractive potential based on scale economies. The optimal city size depends on the balance between its attractive potential and the cost of living that can be approximated by equilibrium land rent and commuting cost. For moderate scale effects optimal population of a city depends negatively on transport costs that are related positively with energy price index. The optimal agricultural density of population can also be constructed. The larger is a land slot per peasant, the higher will be the output from one unit of his labour force applied to this slot. But at the same time, larger farm size results in increase of energy costs, related to land development, collecting the crop and bringing it to the market. In the last 10 years we have observed substantial rise of both food and energy prices at the world stock markets. However, the income of farmers did not grow as fast as food price index. This can shift optimal rural population density to lower level, causing migration to cities (and we observe this tendency globally. Any change in those prices results in suboptimality of existing spatial structures. If changes are slow, the optimal infrastructure can be adjusted by simple migration. If the shocks are high, adaptation may be impossible and shock will persist. This took place in early 1990es in the former USSR, where after transition to world price for oil in domestic markets existing spatial infrastructure became suboptimal and resulted in persistent crisis, leading to deterioration of both industry and agriculture. Russia is the largest country but this is also its problem. Having large resource endowment per capita, it is problematic to build sufficient infrastructure. Russia has too low population
Directory of Open Access Journals (Sweden)
Trine Krogh-Madsen
2017-12-01
Full Text Available In silico cardiac myocyte models present powerful tools for drug safety testing and for predicting phenotypical consequences of ion channel mutations, but their accuracy is sometimes limited. For example, several models describing human ventricular electrophysiology perform poorly when simulating effects of long QT mutations. Model optimization represents one way of obtaining models with stronger predictive power. Using a recent human ventricular myocyte model, we demonstrate that model optimization to clinical long QT data, in conjunction with physiologically-based bounds on intracellular calcium and sodium concentrations, better constrains model parameters. To determine if the model optimized to congenital long QT data better predicts risk of drug-induced long QT arrhythmogenesis, in particular Torsades de Pointes risk, we tested the optimized model against a database of known arrhythmogenic and non-arrhythmogenic ion channel blockers. When doing so, the optimized model provided an improved risk assessment. In particular, we demonstrate an elimination of false-positive outcomes generated by the baseline model, in which simulations of non-torsadogenic drugs, in particular verapamil, predict action potential prolongation. Our results underscore the importance of currents beyond those directly impacted by a drug block in determining torsadogenic risk. Our study also highlights the need for rich data in cardiac myocyte model optimization and substantiates such optimization as a method to generate models with higher accuracy of predictions of drug-induced cardiotoxicity.
A New Method for Global Optimization Based on Stochastic Differential Equations.
1984-12-01
Serie Naranja, n. 204, IINAS-UNAM, Mx ic o D. F. , 1979. [6] A. V. Levy, A. Montalvo, S. G6mez, A. Cald’er6n, ’Topics in global optimi~zation", in: J...FTFOPT aF 455. £ 456. C S7ART SERIES OF TR IAL5 457. C 458. DO 30 IC x 1,M7RIA&. 459. C 46r’. C SET INITIALIZATION IN&EX FOR NOISE GENERATOR 461. C 1 462...Ia iunghezza del passo di integrazione temporale , t k =o+ hi+ h 2+ ... + h kl rk e u ksono due vettori aleatori in n.-dimensioni scelti ii primo da
Kucukgoz, Mehmet; Harmanci, Oztan; Mihcak, Mehmet K.; Venkatesan, Ramarathnam
2005-03-01
In this paper, we propose a novel semi-blind video watermarking scheme, where we use pseudo-random robust semi-global features of video in the three dimensional wavelet transform domain. We design the watermark sequence via solving an optimization problem, such that the features of the mark-embedded video are the quantized versions of the features of the original video. The exact realizations of the algorithmic parameters are chosen pseudo-randomly via a secure pseudo-random number generator, whose seed is the secret key, that is known (resp. unknown) by the embedder and the receiver (resp. by the public). We experimentally show the robustness of our algorithm against several attacks, such as conventional signal processing modifications and adversarial estimation attacks.
Directory of Open Access Journals (Sweden)
Zhongbo Sun
2014-01-01
Full Text Available Two modified three-term type conjugate gradient algorithms which satisfy both the descent condition and the Dai-Liao type conjugacy condition are presented for unconstrained optimization. The first algorithm is a modification of the Hager and Zhang type algorithm in such a way that the search direction is descent and satisfies Dai-Liao’s type conjugacy condition. The second simple three-term type conjugate gradient method can generate sufficient decent directions at every iteration; moreover, this property is independent of the steplength line search. Also, the algorithms could be considered as a modification of the MBFGS method, but with different zk. Under some mild conditions, the given methods are global convergence, which is independent of the Wolfe line search for general functions. The numerical experiments show that the proposed methods are very robust and efficient.
Quantifying global fossil-fuel CO2 emissions: from OCO-2 to optimal observing designs
Ye, X.; Lauvaux, T.; Kort, E. A.; Oda, T.; Feng, S.; Lin, J. C.; Yang, E. G.; Wu, D.; Kuze, A.; Suto, H.; Eldering, A.
2017-12-01
Cities house more than half of the world's population and are responsible for more than 70% of the world anthropogenic CO2 emissions. Therefore, quantifications of emissions from major cities, which are only less than a hundred intense emitting spots across the globe, should allow us to monitor changes in global fossil-fuel CO2 emissions, in an independent, objective way. Satellite platforms provide favorable temporal and spatial coverage to collect urban CO2 data to quantify the anthropogenic contributions to the global carbon budget. We present here the optimal observation design for future NASA's OCO-2 and Japanese GOSAT missions, based on real-data (i.e. OCO-2) experiments and Observing System Simulation Experiments (OSSE's) to address different error components in the urban CO2 budget calculation. We identify the major sources of emission uncertainties for various types of cities with different ecosystems and geographical features, such as urban plumes over flat terrains, accumulated enhancements within basins, and complex weather regimes in coastal areas. Atmospheric transport errors were characterized under various meteorological conditions using the Weather Research and Forecasting (WRF) model at 1-km spatial resolution, coupled to the Open-source Data Inventory for Anthropogenic CO2 (ODIAC) emissions. We propose and discuss the optimized urban sampling strategies to address some difficulties from the seasonality in cloud cover and emissions, vegetation density in and around cities, and address the daytime sampling bias using prescribed diurnal cycles. These factors are combined in pseudo data experiments in which we evaluate the relative impact of uncertainties on inverse estimates of CO2 emissions for cities across latitudinal and climatological zones. We propose here several sampling strategies to minimize the uncertainties in target mode for tracking urban fossil-fuel CO2 emissions over the globe for future satellite missions, such as OCO-3 and future
A Novel Global MPP Tracking of Photovoltaic System based on Whale Optimization Algorithm
Directory of Open Access Journals (Sweden)
Santhan Kumar Cherukuri
2016-11-01
Full Text Available To harvest maximum amount of solar energy and to attain higher efficiency, photovoltaic generation (PVG systems are to be operated at their maximum power point (MPP under both variable climatic and partial shaded condition (PSC. From literature most of conventional MPP tracking (MPPT methods are able to guarantee MPP successfully under uniform shading condition but fails to get global MPP as they may trap at local MPP under PSC, which adversely deteriorates the efficiency of Photovoltaic Generation (PVG system. In this paper a novel MPPT based on Whale Optimization Algorithm (WOA is proposed to analyze analytic modeling of PV system considering both series and shunt resistances for MPP tracking under PSC. The proposed algorithm is tested on 6S, 3S2P and 2S3P Photovoltaic array configurations for different shading patterns and results are presented. To compare the performance, GWO and PSO MPPT algorithms are also simulated and results are also presented. From the results it is noticed that proposed MPPT method is superior to other MPPT methods with reference to accuracy and tracking speed. Article History: Received July 23rd 2016; Received in revised form September 15th 2016; Accepted October 1st 2016; Available online How to Cite This Article: Kumar, C.H.S and Rao, R.S. (2016 A Novel Global MPP Tracking of Photovoltaic System based on Whale Optimization Algorithm. Int. Journal of Renewable Energy Development, 5(3, 225-232. http://dx.doi.org/10.14710/ijred.5.3.225-232
Directory of Open Access Journals (Sweden)
Carlos Pozo
Full Text Available Optimization models in metabolic engineering and systems biology focus typically on optimizing a unique criterion, usually the synthesis rate of a metabolite of interest or the rate of growth. Connectivity and non-linear regulatory effects, however, make it necessary to consider multiple objectives in order to identify useful strategies that balance out different metabolic issues. This is a fundamental aspect, as optimization of maximum yield in a given condition may involve unrealistic values in other key processes. Due to the difficulties associated with detailed non-linear models, analysis using stoichiometric descriptions and linear optimization methods have become rather popular in systems biology. However, despite being useful, these approaches fail in capturing the intrinsic nonlinear nature of the underlying metabolic systems and the regulatory signals involved. Targeting more complex biological systems requires the application of global optimization methods to non-linear representations. In this work we address the multi-objective global optimization of metabolic networks that are described by a special class of models based on the power-law formalism: the generalized mass action (GMA representation. Our goal is to develop global optimization methods capable of efficiently dealing with several biological criteria simultaneously. In order to overcome the numerical difficulties of dealing with multiple criteria in the optimization, we propose a heuristic approach based on the epsilon constraint method that reduces the computational burden of generating a set of Pareto optimal alternatives, each achieving a unique combination of objectives values. To facilitate the post-optimal analysis of these solutions and narrow down their number prior to being tested in the laboratory, we explore the use of Pareto filters that identify the preferred subset of enzymatic profiles. We demonstrate the usefulness of our approach by means of a case study
Pozo, Carlos; Guillén-Gosálbez, Gonzalo; Sorribas, Albert; Jiménez, Laureano
2012-01-01
Optimization models in metabolic engineering and systems biology focus typically on optimizing a unique criterion, usually the synthesis rate of a metabolite of interest or the rate of growth. Connectivity and non-linear regulatory effects, however, make it necessary to consider multiple objectives in order to identify useful strategies that balance out different metabolic issues. This is a fundamental aspect, as optimization of maximum yield in a given condition may involve unrealistic values in other key processes. Due to the difficulties associated with detailed non-linear models, analysis using stoichiometric descriptions and linear optimization methods have become rather popular in systems biology. However, despite being useful, these approaches fail in capturing the intrinsic nonlinear nature of the underlying metabolic systems and the regulatory signals involved. Targeting more complex biological systems requires the application of global optimization methods to non-linear representations. In this work we address the multi-objective global optimization of metabolic networks that are described by a special class of models based on the power-law formalism: the generalized mass action (GMA) representation. Our goal is to develop global optimization methods capable of efficiently dealing with several biological criteria simultaneously. In order to overcome the numerical difficulties of dealing with multiple criteria in the optimization, we propose a heuristic approach based on the epsilon constraint method that reduces the computational burden of generating a set of Pareto optimal alternatives, each achieving a unique combination of objectives values. To facilitate the post-optimal analysis of these solutions and narrow down their number prior to being tested in the laboratory, we explore the use of Pareto filters that identify the preferred subset of enzymatic profiles. We demonstrate the usefulness of our approach by means of a case study that optimizes the
Nacelle Chine Installation Based on Wind-Tunnel Test Using Efficient Global Optimization
Kanazaki, Masahiro; Yokokawa, Yuzuru; Murayama, Mitsuhiro; Ito, Takeshi; Jeong, Shinkyu; Yamamoto, Kazuomi
Design exploration of a nacelle chine installation was carried out. The nacelle chine improves stall performance when deploying multi-element high-lift devices. This study proposes an efficient design process using a Kriging surrogate model to determine the nacelle chine installation point in wind-tunnel tests. The design exploration was conducted in a wind-tunnel using the JAXA high-lift aircraft model at the JAXA Large-scale Low-speed Wind Tunnel. The objective was to maximize the maximum lift. The chine installation points were designed on the engine nacelle in the axial and chord-wise direction, while the geometry of the chine was fixed. In the design process, efficient global optimization (EGO) which includes Kriging model and genetic algorithm (GA) was employed. This method makes it possible both to improve the accuracy of the response surface and to explore the global optimum efficiently. Detailed observations of flowfields using the Particle Image Velocimetry method confirmed the chine effect and design results.
Corzo, Gerald; Solomatine, Dimitri
2007-05-01
Natural phenomena are multistationary and are composed of a number of interacting processes, so one single model handling all processes often suffers from inaccuracies. A solution is to partition data in relation to such processes using the available domain knowledge or expert judgment, to train separate models for each of the processes, and to merge them in a modular model (committee). In this paper a problem of water flow forecast in watershed hydrology is considered where the flow process can be presented as consisting of two subprocesses -- base flow and excess flow, so that these two processes can be separated. Several approaches to data separation techniques are studied. Two case studies with different forecast horizons are considered. Parameters of the algorithms responsible for data partitioning are optimized using genetic algorithms and global pattern search. It was found that modularization of ANN models using domain knowledge makes models more accurate, if compared with a global model trained on the whole data set, especially when forecast horizon (and hence the complexity of the modelled processes) is increased.
Directory of Open Access Journals (Sweden)
Jian-Guo Zheng
2015-01-01
Full Text Available Artificial bee colony (ABC algorithm is a popular swarm intelligence technique inspired by the intelligent foraging behavior of honey bees. However, ABC is good at exploration but poor at exploitation and its convergence speed is also an issue in some cases. To improve the performance of ABC, a novel ABC combined with grenade explosion method (GEM and Cauchy operator, namely, ABCGC, is proposed. GEM is embedded in the onlooker bees’ phase to enhance the exploitation ability and accelerate convergence of ABCGC; meanwhile, Cauchy operator is introduced into the scout bees’ phase to help ABCGC escape from local optimum and further enhance its exploration ability. Two sets of well-known benchmark functions are used to validate the better performance of ABCGC. The experiments confirm that ABCGC is significantly superior to ABC and other competitors; particularly it converges to the global optimum faster in most cases. These results suggest that ABCGC usually achieves a good balance between exploitation and exploration and can effectively serve as an alternative for global optimization.
Local search for optimal global map generation using mid-decadal landsat images
Khatib, L.; Gasch, J.; Morris, Robert; Covington, S.
2007-01-01
NASA and the US Geological Survey (USGS) are seeking to generate a map of the entire globe using Landsat 5 Thematic Mapper (TM) and Landsat 7 Enhanced Thematic Mapper Plus (ETM+) sensor data from the "mid-decadal" period of 2004 through 2006. The global map is comprised of thousands of scene locations and, for each location, tens of different images of varying quality to chose from. Furthermore, it is desirable for images of adjacent scenes be close together in time of acquisition, to avoid obvious discontinuities due to seasonal changes. These characteristics make it desirable to formulate an automated solution to the problem of generating the complete map. This paper formulates a Global Map Generator problem as a Constraint Optimization Problem (GMG-COP) and describes an approach to solving it using local search. Preliminary results of running the algorithm on image data sets are summarized. The results suggest a significant improvement in map quality using constraint-based solutions. Copyright ?? 2007, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
Characterization of PV panel and global optimization of its model parameters using genetic algorithm
International Nuclear Information System (INIS)
Ismail, M.S.; Moghavvemi, M.; Mahlia, T.M.I.
2013-01-01
Highlights: • Genetic Algorithm optimization ability had been utilized to extract parameters of PV panel model. • Effect of solar radiation and temperature variations was taken into account in fitness function evaluation. • We used Matlab-Simulink to simulate operation of the PV-panel to validate results. • Different cases were analyzed to ascertain which of them gives more accurate results. • Accuracy and applicability of this approach to be used as a valuable tool for PV modeling were clearly validated. - Abstract: This paper details an improved modeling technique for a photovoltaic (PV) module; utilizing the optimization ability of a genetic algorithm, with different parameters of the PV module being computed via this approach. The accurate modeling of any PV module is incumbent upon the values of these parameters, as it is imperative in the context of any further studies concerning different PV applications. Simulation, optimization and the design of the hybrid systems that include PV are examples of these applications. The global optimization of the parameters and the applicability for the entire range of the solar radiation and a wide range of temperatures are achievable via this approach. The Manufacturer’s Data Sheet information is used as a basis for the purpose of parameter optimization, with an average absolute error fitness function formulated; and a numerical iterative method used to solve the voltage-current relation of the PV module. The results of single-diode and two-diode models are evaluated in order to ascertain which of them are more accurate. Other cases are also analyzed in this paper for the purpose of comparison. The Matlab–Simulink environment is used to simulate the operation of the PV module, depending on the extracted parameters. The results of the simulation are compared with the Data Sheet information, which is obtained via experimentation in order to validate the reliability of the approach. Three types of PV modules
Directory of Open Access Journals (Sweden)
Ali Wagdy Mohamed
2014-11-01
Full Text Available In this paper, a novel version of Differential Evolution (DE algorithm based on a couple of local search mutation and a restart mechanism for solving global numerical optimization problems over continuous space is presented. The proposed algorithm is named as Restart Differential Evolution algorithm with Local Search Mutation (RDEL. In RDEL, inspired by Particle Swarm Optimization (PSO, a novel local mutation rule based on the position of the best and the worst individuals among the entire population of a particular generation is introduced. The novel local mutation scheme is joined with the basic mutation rule through a linear decreasing function. The proposed local mutation scheme is proven to enhance local search tendency of the basic DE and speed up the convergence. Furthermore, a restart mechanism based on random mutation scheme and a modified Breeder Genetic Algorithm (BGA mutation scheme is combined to avoid stagnation and/or premature convergence. Additionally, an exponent increased crossover probability rule and a uniform scaling factors of DE are introduced to promote the diversity of the population and to improve the search process, respectively. The performance of RDEL is investigated and compared with basic differential evolution, and state-of-the-art parameter adaptive differential evolution variants. It is discovered that the proposed modifications significantly improve the performance of DE in terms of quality of solution, efficiency and robustness.
Directory of Open Access Journals (Sweden)
Zdravko Bazdan
2010-12-01
Full Text Available The aim of this study is to point to the fact that economic diplomacy is a relatively new practice in international economics, specifically the expansion of the occurrence of Intelligence Revolution. The history in global relations shows that without economic diplomacy there is no optimal economic growth and social development. It is important to note that economic diplomacy should be important for our country and the political elite, as well as for the administration of Croatian economic subjects that want to compete in international market economy. Comparative analysis are particularly highlighted by French experience. Therefore, Croatia should copy the practice of those countries that are successful in economic diplomacy. And in the curricula - especially of our economic faculties - we should introduce the course of Economic Diplomacy. It is important to note, that in order to form our optimal model of economic diplomacy which would be headed by the President of Republic of Croatia formula should be based on: Intelligence Security Agency (SOA, Intelligence Service of the Ministry of Foreign Affairs and European Integration, Intelligence Service of the Croatian Chamber of Commerce and the Intelligence Service of the Ministry of Economy, Labor and Entrepreneurship. Described model would consist of intelligence subsystem with at least twelve components.
A global review of freshwater crayfish temperature tolerance, preference, and optimal growth
Westhoff, Jacob T.; Rosenberger, Amanda E.
2016-01-01
Conservation efforts, environmental planning, and management must account for ongoing ecosystem alteration due to a changing climate, introduced species, and shifting land use. This type of management can be facilitated by an understanding of the thermal ecology of aquatic organisms. However, information on thermal ecology for entire taxonomic groups is rarely compiled or summarized, and reviews of the science can facilitate its advancement. Crayfish are one of the most globally threatened taxa, and ongoing declines and extirpation could have serious consequences on aquatic ecosystem function due to their significant biomass and ecosystem roles. Our goal was to review the literature on thermal ecology for freshwater crayfish worldwide, with emphasis on studies that estimated temperature tolerance, temperature preference, or optimal growth. We also explored relationships between temperature metrics and species distributions. We located 56 studies containing information for at least one of those three metrics, which covered approximately 6 % of extant crayfish species worldwide. Information on one or more metrics existed for all 3 genera of Astacidae, 4 of the 12 genera of Cambaridae, and 3 of the 15 genera of Parastacidae. Investigations employed numerous methodological approaches for estimating these parameters, which restricts comparisons among and within species. The only statistically significant relationship we observed between a temperature metric and species range was a negative linear relationship between absolute latitude and optimal growth temperature. We recommend expansion of studies examining the thermal ecology of freshwater crayfish and identify and discuss methodological approaches that can improve standardization and comparability among studies.
3D prostate TRUS segmentation using globally optimized volume-preserving prior.
Qiu, Wu; Rajchl, Martin; Guo, Fumin; Sun, Yue; Ukwatta, Eranga; Fenster, Aaron; Yuan, Jing
2014-01-01
An efficient and accurate segmentation of 3D transrectal ultrasound (TRUS) images plays an important role in the planning and treatment of the practical 3D TRUS guided prostate biopsy. However, a meaningful segmentation of 3D TRUS images tends to suffer from US speckles, shadowing and missing edges etc, which make it a challenging task to delineate the correct prostate boundaries. In this paper, we propose a novel convex optimization based approach to extracting the prostate surface from the given 3D TRUS image, while preserving a new global volume-size prior. We, especially, study the proposed combinatorial optimization problem by convex relaxation and introduce its dual continuous max-flow formulation with the new bounded flow conservation constraint, which results in an efficient numerical solver implemented on GPUs. Experimental results using 12 patient 3D TRUS images show that the proposed approach while preserving the volume-size prior yielded a mean DSC of 89.5% +/- 2.4%, a MAD of 1.4 +/- 0.6 mm, a MAXD of 5.2 +/- 3.2 mm, and a VD of 7.5% +/- 6.2% in - 1 minute, deomonstrating the advantages of both accuracy and efficiency. In addition, the low standard deviation of the segmentation accuracy shows a good reliability of the proposed approach.
Directory of Open Access Journals (Sweden)
Jarmo Nurmi
2017-05-01
Full Text Available This paper addresses the energy-inefficiency problem of four-degrees-of-freedom (4-DOF hydraulic manipulators through redundancy resolution in robotic closed-loop controlled applications. Because conventional methods typically are local and have poor performance for resolving redundancy with respect to minimum hydraulic energy consumption, global energy-optimal redundancy resolution is proposed at the valve-controlled actuator and hydraulic power system interaction level. The energy consumption of the widely popular valve-controlled load-sensing (LS and constant-pressure (CP systems is effectively minimised through cost functions formulated in a discrete-time dynamic programming (DP approach with minimum state representation. A prescribed end-effector path and important actuator constraints at the position, velocity and acceleration levels are also satisfied in the solution. Extensive field experiments performed on a forestry hydraulic manipulator demonstrate the performance of the proposed solution. Approximately 15–30% greater hydraulic energy consumption was observed with the conventional methods in the LS and CP systems. These results encourage energy-optimal redundancy resolution in future robotic applications of hydraulic manipulators.
Development of a fuzzy optimization model, supporting global warming decision-making
International Nuclear Information System (INIS)
Leimbach, M.
1996-01-01
An increasing number of models have been developed to support global warming response policies. The model constructors are facing a lot of uncertainties which limit the evidence of these models. The support of climate policy decision-making is only possible in a semi-quantitative way, as presented by a Fuzzy model. The model design is based on an optimization approach, integrated in a bounded risk decision-making framework. Given some regional emission-related and impact-related restrictions, optimal emission paths can be calculated. The focus is not only on carbon dioxide but on other greenhouse gases too. In the paper, the components of the model will be described. Cost coefficients, emission boundaries and impact boundaries are represented as Fuzzy parameters. The Fuzzy model will be transformed into a computational one by using an approach of Rommelfanger. In the second part, some problems of applying the model to computations will be discussed. This includes discussions on the data situation and the presentation, as well as interpretation of results of sensitivity analyses. The advantage of the Fuzzy approach is that the requirements regarding data precision are not so strong. Hence, the effort for data acquisition can be reduced and computations can be started earlier. 9 figs., 3 tabs., 17 refs., 1 appendix
Energy Technology Data Exchange (ETDEWEB)
Shen, Bo [ORNL; Abdelaziz, Omar [ORNL; Shrestha, Som S [ORNL
2017-01-01
Oak Ridge National laboratory (ORNL) recently conducted extensive laboratory, drop-in investigations for lower Global Warming Potential (GWP) refrigerants to replace R-22 and R-410A. ORNL studied propane, DR-3, ARM-20B, N-20B and R-444B as lower GWP refrigerant replacement for R-22 in a mini-split room air conditioner (RAC) originally designed for R-22; and, R-32, DR-55, ARM-71A, and L41-2, in a mini-split RAC designed for R-410A. We obtained laboratory testing results with very good energy balance and nominal measurement uncertainty. Drop-in studies are not enough to judge the overall performance of the alternative refrigerants since their thermodynamic and transport properties might favor different heat exchanger configurations, e.g. cross-flow, counter flow, etc. This study compares optimized performances of individual refrigerants using a physics-based system model tools. The DOE/ORNL Heat Pump Design Model (HPDM) was used to model the mini-split RACs by inputting detailed heat exchangers geometries, compressor displacement and efficiencies as well as other relevant system components. The RAC models were calibrated against the lab data for each individual refrigerant. The calibrated models were then used to conduct a design optimization for the cooling performance by varying the compressor displacement to match the required capacity, and changing the number of circuits, refrigerant flow direction, tube diameters, air flow rates in the condenser and evaporator at 100% and 50% cooling capacities. This paper compares the optimized performance results for all alternative refrigerants and highlights best candidates for R-22 and R-410A replacement.
Vaziri Yazdi Pin, Mohammad
practices. Single criterion optimization algorithms using mathematical programming for globally optimal solutions have been developed for three objectives of cost, reliability, and the social/environmental impacts. Additional algorithms for inclusions of upgrade and optimal load assignment possibilities have been developed. Algorithms have been developed to handle the expansion as a multiobjective decision process. Typical data from both major investor owned and major municipal utilities operating in California USA, have been utilized to implement and test the algorithms on practical test cases. Results of the case studies and associated analyses indicate that the developed algorithms also perform efficiently in solving the multistage and multiobjective expansion problem.
The bounds of feasible space on constrained nonconvex quadratic programming
Zhu, Jinghao
2008-03-01
This paper presents a method to estimate the bounds of the radius of the feasible space for a class of constrained nonconvex quadratic programmingsE Results show that one may compute a bound of the radius of the feasible space by a linear programming which is known to be a P-problem [N. Karmarkar, A new polynomial-time algorithm for linear programming, Combinatorica 4 (1984) 373-395]. It is proposed that one applies this method for using the canonical dual transformation [D.Y. Gao, Canonical duality theory and solutions to constrained nonconvex quadratic programming, J. Global Optimization 29 (2004) 377-399] for solving a standard quadratic programming problem.
Energy Technology Data Exchange (ETDEWEB)
Portnoy, David, E-mail: david.portnoy@jhuapl.edu [Johns Hopkins University Applied Physics Laboratory, 11100 Johns Hopkins Road, Laurel, MD 20723 (United States); Feuerbach, Robert; Heimberg, Jennifer [Johns Hopkins University Applied Physics Laboratory, 11100 Johns Hopkins Road, Laurel, MD 20723 (United States)
2011-10-01
Today there is a tremendous amount of interest in systems that can detect radiological or nuclear threats. Many of these systems operate in extremely high throughput situations where delays caused by false alarms can have a significant negative impact. Thus, calculating the tradeoff between detection rates and false alarm rates is critical for their successful operation. Receiver operating characteristic (ROC) curves have long been used to depict this tradeoff. The methodology was first developed in the field of signal detection. In recent years it has been used increasingly in machine learning and data mining applications. It follows that this methodology could be applied to radiological/nuclear threat detection systems. However many of these systems do not fit into the classic principles of statistical detection theory because they tend to lack tractable likelihood functions and have many parameters, which, in general, do not have a one-to-one correspondence with the detection classes. This work proposes a strategy to overcome these problems by empirically finding parameter values that maximize the probability of detection for a selected number of probabilities of false alarm. To find these parameter values a statistical global optimization technique that seeks to estimate portions of a ROC curve is proposed. The optimization combines elements of simulated annealing with elements of genetic algorithms. Genetic algorithms were chosen because they can reduce the risk of getting stuck in local minima. However classic genetic algorithms operate on arrays of Booleans values or bit strings, so simulated annealing is employed to perform mutation in the genetic algorithm. The presented initial results were generated using an isotope identification algorithm developed at Johns Hopkins University Applied Physics Laboratory. The algorithm has 12 parameters: 4 real-valued and 8 Boolean. A simulated dataset was used for the optimization study; the 'threat' set of
International Nuclear Information System (INIS)
Portnoy, David; Feuerbach, Robert; Heimberg, Jennifer
2011-01-01
Today there is a tremendous amount of interest in systems that can detect radiological or nuclear threats. Many of these systems operate in extremely high throughput situations where delays caused by false alarms can have a significant negative impact. Thus, calculating the tradeoff between detection rates and false alarm rates is critical for their successful operation. Receiver operating characteristic (ROC) curves have long been used to depict this tradeoff. The methodology was first developed in the field of signal detection. In recent years it has been used increasingly in machine learning and data mining applications. It follows that this methodology could be applied to radiological/nuclear threat detection systems. However many of these systems do not fit into the classic principles of statistical detection theory because they tend to lack tractable likelihood functions and have many parameters, which, in general, do not have a one-to-one correspondence with the detection classes. This work proposes a strategy to overcome these problems by empirically finding parameter values that maximize the probability of detection for a selected number of probabilities of false alarm. To find these parameter values a statistical global optimization technique that seeks to estimate portions of a ROC curve is proposed. The optimization combines elements of simulated annealing with elements of genetic algorithms. Genetic algorithms were chosen because they can reduce the risk of getting stuck in local minima. However classic genetic algorithms operate on arrays of Booleans values or bit strings, so simulated annealing is employed to perform mutation in the genetic algorithm. The presented initial results were generated using an isotope identification algorithm developed at Johns Hopkins University Applied Physics Laboratory. The algorithm has 12 parameters: 4 real-valued and 8 Boolean. A simulated dataset was used for the optimization study; the 'threat' set of spectra
Portnoy, David; Feuerbach, Robert; Heimberg, Jennifer
2011-10-01
Today there is a tremendous amount of interest in systems that can detect radiological or nuclear threats. Many of these systems operate in extremely high throughput situations where delays caused by false alarms can have a significant negative impact. Thus, calculating the tradeoff between detection rates and false alarm rates is critical for their successful operation. Receiver operating characteristic (ROC) curves have long been used to depict this tradeoff. The methodology was first developed in the field of signal detection. In recent years it has been used increasingly in machine learning and data mining applications. It follows that this methodology could be applied to radiological/nuclear threat detection systems. However many of these systems do not fit into the classic principles of statistical detection theory because they tend to lack tractable likelihood functions and have many parameters, which, in general, do not have a one-to-one correspondence with the detection classes. This work proposes a strategy to overcome these problems by empirically finding parameter values that maximize the probability of detection for a selected number of probabilities of false alarm. To find these parameter values a statistical global optimization technique that seeks to estimate portions of a ROC curve is proposed. The optimization combines elements of simulated annealing with elements of genetic algorithms. Genetic algorithms were chosen because they can reduce the risk of getting stuck in local minima. However classic genetic algorithms operate on arrays of Booleans values or bit strings, so simulated annealing is employed to perform mutation in the genetic algorithm. The presented initial results were generated using an isotope identification algorithm developed at Johns Hopkins University Applied Physics Laboratory. The algorithm has 12 parameters: 4 real-valued and 8 Boolean. A simulated dataset was used for the optimization study; the "threat" set of spectra
Climate, Agriculture, Energy and the Optimal Allocation of Global Land Use
Steinbuks, J.; Hertel, T. W.
2011-12-01
The allocation of the world's land resources over the course of the next century has become a pressing research question. Continuing population increases, improving, land-intensive diets amongst the poorest populations in the world, increasing production of biofuels and rapid urbanization in developing countries are all competing for land even as the world looks to land resources to supply more environmental services. The latter include biodiversity and natural lands, as well as forests and grasslands devoted to carbon sequestration. And all of this is taking place in the context of faster than expected climate change which is altering the biophysical environment for land-related activities. The goal of the paper is to determine the optimal profile for global land use in the context of growing commercial demands for food and forest products, increasing non-market demands for ecosystem services, and more stringent GHG mitigation targets. We then seek to assess how the uncertainty associated with the underlying biophysical and economic processes influences this optimal profile of land use, in light of potential irreversibility in these decisions. We develop a dynamic long-run, forward-looking partial equilibrium framework in which the societal objective function being maximized places value on food production, liquid fuels (including biofuels), timber production, forest carbon and biodiversity. Given the importance of land-based emissions to any GHG mitigation strategy, as well as the potential impacts of climate change itself on the productivity of land in agriculture, forestry and ecosystem services, we aim to identify the optimal allocation of the world's land resources, over the course of the next century, in the face of alternative GHG constraints. The forestry sector is characterized by multiple forest vintages which add considerable computational complexity in the context of this dynamic analysis. In order to solve this model efficiently, we have employed the
Keeling, Charles D.; Piper, S. C.
1998-01-01
Our original proposal called for improved modeling of the terrestrial biospheric carbon cycle, specifically using biome-specific process models to account for both the energy and water budgets of plant growth, to facilitate investigations into recent changes in global atmospheric CO2 abundance and regional distribution. The carbon fluxes predicted by these models were to be incorporated into a global model of CO2 transport to establish large-scale regional fluxes of CO2 to and from the terrestrial biosphere subject to constraints imposed by direct measurements of atmospheric CO2 and its 13C/12C isotopic ratio. Our work was coordinated with a NASA project (NASA NAGW-3151) at the University of Montana under the direction of Steven Running, and was partially funded by the Electric Power Research Institute. The primary objective of this project was to develop and test the Biome-BGC model, a global biological process model with a daily time step which simulates the water, energy and carbon budgets of plant growth. The primary product, the unique global gridded daily land temperature, and the precipitation data set which was used to drive the process model is described. The Biome-BGC model was tested by comparison with a simpler biological model driven by satellite-derived (NDVI) Normalized Difference Vegetation Index and (PAR) Photosynthetically Active Radiation data and by comparison with atmospheric CO2 observations. The simple NDVI model is also described. To facilitate the comparison with atmospheric CO2 observations, a three-dimensional atmospheric transport model was used to produce predictions of atmospheric CO2 variations given CO2 fluxes owing to (NPP) Net Primary Productivity and heterotrophic respiration that were produced by the Biome-BGC model and by the NDVI model. The transport model that we used in this project, and errors associated with transport simulations, were characterized by a comparison of 12 transport models.
International Nuclear Information System (INIS)
Ioannou, Lawrence M.; Travaglione, Benjamin C.
2006-01-01
We focus on determining the separability of an unknown bipartite quantum state ρ by invoking a sufficiently large subset of all possible entanglement witnesses given the expected value of each element of a set of mutually orthogonal observables. We review the concept of an entanglement witness from the geometrical point of view and use this geometry to show that the set of separable states is not a polytope and to characterize the class of entanglement witnesses (observables) that detect entangled states on opposite sides of the set of separable states. All this serves to motivate a classical algorithm which, given the expected values of a subset of an orthogonal basis of observables of an otherwise unknown quantum state, searches for an entanglement witness in the span of the subset of observables. The idea of such an algorithm, which is an efficient reduction of the quantum separability problem to a global optimization problem, was introduced by [Ioannou et al., Phys. Rev. A 70, 060303(R)], where it was shown to be an improvement on the naive approach for the quantum separability problem (exhaustive search for a decomposition of the given state into a convex combination of separable states). The last section of the paper discusses in more generality such algorithms, which, in our case, assume a subroutine that computes the global maximum of a real function of several variables. Despite this, we anticipate that such algorithms will perform sufficiently well on small instances that they will render a feasible test for separability in some cases of interest (e.g., in 3x3 dimensional systems)
Pozo, Carlos; Marín-Sanguino, Alberto; Alves, Rui; Guillén-Gosálbez, Gonzalo; Jiménez, Laureano; Sorribas, Albert
2011-08-25
Design of newly engineered microbial strains for biotechnological purposes would greatly benefit from the development of realistic mathematical models for the processes to be optimized. Such models can then be analyzed and, with the development and application of appropriate optimization techniques, one could identify the modifications that need to be made to the organism in order to achieve the desired biotechnological goal. As appropriate models to perform such an analysis are necessarily non-linear and typically non-convex, finding their global optimum is a challenging task. Canonical modeling techniques, such as Generalized Mass Action (GMA) models based on the power-law formalism, offer a possible solution to this problem because they have a mathematical structure that enables the development of specific algorithms for global optimization. Based on the GMA canonical representation, we have developed in previous works a highly efficient optimization algorithm and a set of related strategies for understanding the evolution of adaptive responses in cellular metabolism. Here, we explore the possibility of recasting kinetic non-linear models into an equivalent GMA model, so that global optimization on the recast GMA model can be performed. With this technique, optimization is greatly facilitated and the results are transposable to the original non-linear problem. This procedure is straightforward for a particular class of non-linear models known as Saturable and Cooperative (SC) models that extend the power-law formalism to deal with saturation and cooperativity. Our results show that recasting non-linear kinetic models into GMA models is indeed an appropriate strategy that helps overcoming some of the numerical difficulties that arise during the global optimization task.
Directory of Open Access Journals (Sweden)
Sorribas Albert
2011-08-01
Full Text Available Abstract Background Design of newly engineered microbial strains for biotechnological purposes would greatly benefit from the development of realistic mathematical models for the processes to be optimized. Such models can then be analyzed and, with the development and application of appropriate optimization techniques, one could identify the modifications that need to be made to the organism in order to achieve the desired biotechnological goal. As appropriate models to perform such an analysis are necessarily non-linear and typically non-convex, finding their global optimum is a challenging task. Canonical modeling techniques, such as Generalized Mass Action (GMA models based on the power-law formalism, offer a possible solution to this problem because they have a mathematical structure that enables the development of specific algorithms for global optimization. Results Based on the GMA canonical representation, we have developed in previous works a highly efficient optimization algorithm and a set of related strategies for understanding the evolution of adaptive responses in cellular metabolism. Here, we explore the possibility of recasting kinetic non-linear models into an equivalent GMA model, so that global optimization on the recast GMA model can be performed. With this technique, optimization is greatly facilitated and the results are transposable to the original non-linear problem. This procedure is straightforward for a particular class of non-linear models known as Saturable and Cooperative (SC models that extend the power-law formalism to deal with saturation and cooperativity. Conclusions Our results show that recasting non-linear kinetic models into GMA models is indeed an appropriate strategy that helps overcoming some of the numerical difficulties that arise during the global optimization task.
International Nuclear Information System (INIS)
Voyant, Cyril; Muselli, Marc; Paoli, Christophe; Nivet, Marie-Laure
2011-01-01
This paper presents an application of Artificial Neural Networks (ANNs) to predict daily solar radiation. We look at the Multi-Layer Perceptron (MLP) network which is the most used of ANNs architectures. In previous studies, we have developed an ad-hoc time series preprocessing and optimized a MLP with endogenous inputs in order to forecast the solar radiation on a horizontal surface. We propose in this paper to study the contribution of exogenous meteorological data (multivariate method) as time series to our optimized MLP and compare with different forecasting methods: a naive forecaster (persistence), ARIMA reference predictor, an ANN with preprocessing using only endogenous inputs (univariate method) and an ANN with preprocessing using endogenous and exogenous inputs. The use of exogenous data generates an nRMSE decrease between 0.5% and 1% for two stations during 2006 and 2007 (Corsica Island, France). The prediction results are also relevant for the concrete case of a tilted PV wall (1.175 kWp). The addition of endogenous and exogenous data allows a 1% decrease of the nRMSE over a 6 months-cloudy period for the power production. While the use of exogenous data shows an interest in winter, endogenous data as inputs on a preprocessed ANN seem sufficient in summer. -- Research highlights: → Use of exogenous data as ANN inputs to forecast horizontal daily global irradiation data. → New methodology allowing to choice the adequate exogenous data - a systematic method comparing endogenous and exogenous data. → Different referenced mathematical predictors allows to conclude about the pertinence of the proposed methodology.
Optimal estimation of regional N2O emissions using a three-dimensional global model
Huang, J.; Golombek, A.; Prinn, R.
2004-12-01
In this study, we use the MATCH (Model of Atmospheric Transport and Chemistry) model and Kalman filtering techniques to optimally estimate N2O emissions from seven source regions around the globe. The MATCH model was used with NCEP assimilated winds at T62 resolution (192 longitude by 94 latitude surface grid, and 28 vertical levels) from July 1st 1996 to December 31st 2000. The average concentrations of N2O in the lowest four layers of the model were then compared with the monthly mean observations from six national/global networks (AGAGE, CMDL (HATS), CMDL (CCGG), CSIRO, CSIR and NIES), at 48 surface sites. A 12-month-running-mean smoother was applied to both the model results and the observations, due to the fact that the model was not able to reproduce the very small observed seasonal variations. The Kalman filter was then used to solve for the time-averaged regional emissions of N2O for January 1st 1997 to June 30th 2000. The inversions assume that the model stratospheric destruction rates, which lead to a global N2O lifetime of 130 years, are correct. It also assumes normalized emission spatial distributions from each region based on previous studies. We conclude that the global N2O emission flux is about 16.2 TgN/yr, with {34.9±1.7%} from South America and Africa, {34.6±1.5%} from South Asia, {13.9±1.5%} from China/Japan/South East Asia, {8.0±1.9%} from all oceans, {6.4±1.1%} from North America and North and West Asia, {2.6±0.4%} from Europe, and {0.9±0.7%} from New Zealand and Australia. The errors here include the measurement standard deviation, calibration differences among the six groups, grid volume/measurement site mis-match errors estimated from the model, and a procedure to account approximately for the modeling errors.
Kleijnen, Jack P.C.; van Beers, W.C.M.; van Nieuwenhuyse, I.
2010-01-01
This paper uses a sequentialized experimental design to select simulation input com- binations for global optimization, based on Kriging (also called Gaussian process or spatial correlation modeling); this Kriging is used to analyze the input/output data of the simulation model (computer code). This
Directory of Open Access Journals (Sweden)
Tulio Rosembuj
2006-12-01
Full Text Available There is no singular globalization, nor is the result of an individual agent. We could start by saying that global action has different angles and subjects who perform it are different, as well as its objectives. The global is an invisible invasion of materials and immediate effects.
Tulio Rosembuj
2006-01-01
There is no singular globalization, nor is the result of an individual agent. We could start by saying that global action has different angles and subjects who perform it are different, as well as its objectives. The global is an invisible invasion of materials and immediate effects.
Protein structure modeling and refinement by global optimization in CASP12.
Hong, Seung Hwan; Joung, InSuk; Flores-Canales, Jose C; Manavalan, Balachandran; Cheng, Qianyi; Heo, Seungryong; Kim, Jong Yun; Lee, Sun Young; Nam, Mikyung; Joo, Keehyoung; Lee, In-Ho; Lee, Sung Jong; Lee, Jooyoung
2018-03-01
For protein structure modeling in the CASP12 experiment, we have developed a new protocol based on our previous CASP11 approach. The global optimization method of conformational space annealing (CSA) was applied to 3 stages of modeling: multiple sequence-structure alignment, three-dimensional (3D) chain building, and side-chain re-modeling. For better template selection and model selection, we updated our model quality assessment (QA) method with the newly developed SVMQA (support vector machine for quality assessment). For 3D chain building, we updated our energy function by including restraints generated from predicted residue-residue contacts. New energy terms for the predicted secondary structure and predicted solvent accessible surface area were also introduced. For difficult targets, we proposed a new method, LEEab, where the template term played a less significant role than it did in LEE, complemented by increased contributions from other terms such as the predicted contact term. For TBM (template-based modeling) targets, LEE performed better than LEEab, but for FM targets, LEEab was better. For model refinement, we modified our CASP11 molecular dynamics (MD) based protocol by using explicit solvents and tuning down restraint weights. Refinement results from MD simulations that used a new augmented statistical energy term in the force field were quite promising. Finally, when using inaccurate information (such as the predicted contacts), it was important to use the Lorentzian function for which the maximal penalty arising from wrong information is always bounded. © 2017 Wiley Periodicals, Inc.
Two-stage collaborative global optimization design model of the CHPG microgrid
Liao, Qingfen; Xu, Yeyan; Tang, Fei; Peng, Sicheng; Yang, Zheng
2017-06-01
With the continuous developing of technology and reducing of investment costs, renewable energy proportion in the power grid is becoming higher and higher because of the clean and environmental characteristics, which may need more larger-capacity energy storage devices, increasing the cost. A two-stage collaborative global optimization design model of the combined-heat-power-and-gas (abbreviated as CHPG) microgrid is proposed in this paper, to minimize the cost by using virtual storage without extending the existing storage system. P2G technology is used as virtual multi-energy storage in CHPG, which can coordinate the operation of electric energy network and natural gas network at the same time. Demand response is also one kind of good virtual storage, including economic guide for the DGs and heat pumps in demand side and priority scheduling of controllable loads. Two kinds of storage will coordinate to smooth the high-frequency fluctuations and low-frequency fluctuations of renewable energy respectively, and achieve a lower-cost operation scheme simultaneously. Finally, the feasibility and superiority of proposed design model is proved in a simulation of a CHPG microgrid.
International Nuclear Information System (INIS)
Yang, Jian; Cong, Weijian; Fan, Jingfan; Liu, Yue; Wang, Yongtian; Chen, Yang
2014-01-01
The clinical value of the 3D reconstruction of a coronary artery is important for the diagnosis and intervention of cardiovascular diseases. This work proposes a method based on a deformable model for reconstructing coronary arteries from two monoplane angiographic images acquired from different angles. First, an external force back-projective composition model is developed to determine the external force, for which the force distributions in different views are back-projected to the 3D space and composited in the same coordinate system based on the perspective projection principle of x-ray imaging. The elasticity and bending forces are composited as an internal force to maintain the smoothness of the deformable curve. Second, the deformable curve evolves rapidly toward the true vascular centerlines in 3D space and angiographic images under the combination of internal and external forces. Third, densely matched correspondence among vessel centerlines is constructed using a curve alignment method. The bundle adjustment method is then utilized for the global optimization of the projection parameters and the 3D structures. The proposed method is validated on phantom data and routine angiographic images with consideration for space and re-projection image errors. Experimental results demonstrate the effectiveness and robustness of the proposed method for the reconstruction of coronary arteries from two monoplane angiographic images. The proposed method can achieve a mean space error of 0.564 mm and a mean re-projection error of 0.349 mm. (paper)