A convex optimization approach for solving large scale linear systems
Directory of Open Access Journals (Sweden)
Debora Cores
2017-01-01
Full Text Available The well-known Conjugate Gradient (CG method minimizes a strictly convex quadratic function for solving large-scale linear system of equations when the coefficient matrix is symmetric and positive definite. In this work we present and analyze a non-quadratic convex function for solving any large-scale linear system of equations regardless of the characteristics of the coefficient matrix. For finding the global minimizers, of this new convex function, any low-cost iterative optimization technique could be applied. In particular, we propose to use the low-cost globally convergent Spectral Projected Gradient (SPG method, which allow us to extend this optimization approach for solving consistent square and rectangular linear system, as well as linear feasibility problem, with and without convex constraints and with and without preconditioning strategies. Our numerical results indicate that the new scheme outperforms state-of-the-art iterative techniques for solving linear systems when the symmetric part of the coefficient matrix is indefinite, and also for solving linear feasibility problems.
Planning under uncertainty solving large-scale stochastic linear programs
Energy Technology Data Exchange (ETDEWEB)
Infanger, G. [Stanford Univ., CA (United States). Dept. of Operations Research]|[Technische Univ., Vienna (Austria). Inst. fuer Energiewirtschaft
1992-12-01
For many practical problems, solutions obtained from deterministic models are unsatisfactory because they fail to hedge against certain contingencies that may occur in the future. Stochastic models address this shortcoming, but up to recently seemed to be intractable due to their size. Recent advances both in solution algorithms and in computer technology now allow us to solve important and general classes of practical stochastic problems. We show how large-scale stochastic linear programs can be efficiently solved by combining classical decomposition and Monte Carlo (importance) sampling techniques. We discuss the methodology for solving two-stage stochastic linear programs with recourse, present numerical results of large problems with numerous stochastic parameters, show how to efficiently implement the methodology on a parallel multi-computer and derive the theory for solving a general class of multi-stage problems with dependency of the stochastic parameters within a stage and between different stages.
Solving large mixed linear models using preconditioned conjugate gradient iteration.
Strandén, I; Lidauer, M
1999-12-01
Continuous evaluation of dairy cattle with a random regression test-day model requires a fast solving method and algorithm. A new computing technique feasible in Jacobi and conjugate gradient based iterative methods using iteration on data is presented. In the new computing technique, the calculations in multiplication of a vector by a matrix were recorded to three steps instead of the commonly used two steps. The three-step method was implemented in a general mixed linear model program that used preconditioned conjugate gradient iteration. Performance of this program in comparison to other general solving programs was assessed via estimation of breeding values using univariate, multivariate, and random regression test-day models. Central processing unit time per iteration with the new three-step technique was, at best, one-third that needed with the old technique. Performance was best with the test-day model, which was the largest and most complex model used. The new program did well in comparison to other general software. Programs keeping the mixed model equations in random access memory required at least 20 and 435% more time to solve the univariate and multivariate animal models, respectively. Computations of the second best iteration on data took approximately three and five times longer for the animal and test-day models, respectively, than did the new program. Good performance was due to fast computing time per iteration and quick convergence to the final solutions. Use of preconditioned conjugate gradient based methods in solving large breeding value problems is supported by our findings.
Krylov subspace methods for solving large unsymmetric linear systems
International Nuclear Information System (INIS)
Saad, Y.
1981-01-01
Some algorithms based upon a projection process onto the Krylov subspace K/sub m/ = Span(r 0 , Ar 0 ,...,A/sup m/-1r 0 ) are developed, generalizing the method of conjugate gradients to unsymmetric systems. These methods are extensions of Arnoldi's algorithm for solving eigenvalue problems. The convergence is analyzed in terms of the distance of the solution to the subspace K/sub m/ and some error bounds are established showing, in particular, a similarity with the conjugate gradient method (for symmetric matrices) when the eigenvalues are real. Several numerical experiments are described and discussed
Solving large linear systems in an implicit thermohaline ocean model
de Niet, Arie Christiaan
2007-01-01
The climate on earth is largely determined by the global ocean circulation. Hence it is important to predict how the flow will react to perturbation by for example melting icecaps. To answer questions about the stability of the global ocean flow, a computer model has been developed that is able to
International Nuclear Information System (INIS)
Gene Golub; Kwok Ko
2009-01-01
The solutions of sparse eigenvalue problems and linear systems constitute one of the key computational kernels in the discretization of partial differential equations for the modeling of linear accelerators. The computational challenges faced by existing techniques for solving those sparse eigenvalue problems and linear systems call for continuing research to improve on the algorithms so that ever increasing problem size as required by the physics application can be tackled. Under the support of this award, the filter algorithm for solving large sparse eigenvalue problems was developed at Stanford to address the computational difficulties in the previous methods with the goal to enable accelerator simulations on then the world largest unclassified supercomputer at NERSC for this class of problems. Specifically, a new method, the Hemitian skew-Hemitian splitting method, was proposed and researched as an improved method for solving linear systems with non-Hermitian positive definite and semidefinite matrices.
Directory of Open Access Journals (Sweden)
Jen-Yuan Chen
2014-01-01
Full Text Available Continuing from the works of Li et al. (2014, Li (2007, and Kincaid et al. (2000, we present more generalizations and modifications of iterative methods for solving large sparse symmetric and nonsymmetric indefinite systems of linear equations. We discuss a variety of iterative methods such as GMRES, MGMRES, MINRES, LQ-MINRES, QR MINRES, MMINRES, MGRES, and others.
Dobbs, David E.
2013-01-01
A direct method is given for solving first-order linear recurrences with constant coefficients. The limiting value of that solution is studied as "n to infinity." This classroom note could serve as enrichment material for the typical introductory course on discrete mathematics that follows a calculus course.
Using a grid platform for solving large sparse linear systems over GF(2)
Kleinjung , Thorsten; Nussbaum , Lucas; Thomé , Emmanuel
2010-01-01
International audience; In Fall 2009, the final step of the factorization of rsa768 was carried out on several clusters of the Grid'5000 platform, leading to a new record in integer factorization. This step involves solving a huge sparse linear system defined over the binary field GF(2). This article aims at describing the algorithm used, the difficulties encountered, and the methodology which led to success. In particular, we illustrate how our use of the block Wiedemann algorithm led to a m...
An implicit iterative scheme for solving large systems of linear equations
International Nuclear Information System (INIS)
Barry, J.M.; Pollard, J.P.
1986-12-01
An implicit iterative scheme for the solution of large systems of linear equations arising from neutron diffusion studies is presented. The method is applied to three-dimensional reactor studies and its performance is compared with alternative iterative approaches
Energy Technology Data Exchange (ETDEWEB)
Moryakov, A. V., E-mail: sailor@orc.ru [National Research Centre Kurchatov Institute (Russian Federation)
2016-12-15
An algorithm for solving the linear Cauchy problem for large systems of ordinary differential equations is presented. The algorithm for systems of first-order differential equations is implemented in the EDELWEISS code with the possibility of parallel computations on supercomputers employing the MPI (Message Passing Interface) standard for the data exchange between parallel processes. The solution is represented by a series of orthogonal polynomials on the interval [0, 1]. The algorithm is characterized by simplicity and the possibility to solve nonlinear problems with a correction of the operator in accordance with the solution obtained in the previous iterative process.
Efficient Solving of Large Non-linear Arithmetic Constraint Systems with Complex Boolean Structure
Czech Academy of Sciences Publication Activity Database
Fränzle, M.; Herde, C.; Teige, T.; Ratschan, Stefan; Schubert, T.
2007-01-01
Roč. 1, - (2007), s. 209-236 ISSN 1574-0617 Grant - others:AVACS(DE) SFB/TR 14 Institutional research plan: CEZ:AV0Z10300504 Keywords : interval-based arithmetic constraint solving * SAT modulo theories Subject RIV: BA - General Mathematics
Solving Linear Differential Equations
Nguyen, K.A.; Put, M. van der
2010-01-01
The theme of this paper is to 'solve' an absolutely irreducible differential module explicitly in terms of modules of lower dimension and finite extensions of the differential field K. Representations of semi-simple Lie algebras and differential Galo is theory are the main tools. The results extend
Solving sparse linear least squares problems on some supercomputers by using large dense blocks
DEFF Research Database (Denmark)
Hansen, Per Christian; Ostromsky, T; Sameh, A
1997-01-01
technique is preferable to sparse matrix technique when the matrices are not large, because the high computational speed compensates fully the disadvantages of using more arithmetic operations and more storage. For very large matrices the computations must be organized as a sequence of tasks in each......Efficient subroutines for dense matrix computations have recently been developed and are available on many high-speed computers. On some computers the speed of many dense matrix operations is near to the peak-performance. For sparse matrices storage and operations can be saved by operating only...... and storing only nonzero elements. However, the price is a great degradation of the speed of computations on supercomputers (due to the use of indirect addresses, to the need to insert new nonzeros in the sparse storage scheme, to the lack of data locality, etc.). On many high-speed computers a dense matrix...
Nguyen, Duc T.; Mohammed, Ahmed Ali; Kadiam, Subhash
2010-01-01
Solving large (and sparse) system of simultaneous linear equations has been (and continues to be) a major challenging problem for many real-world engineering/science applications [1-2]. For many practical/large-scale problems, the sparse, Symmetrical and Positive Definite (SPD) system of linear equations can be conveniently represented in matrix notation as [A] {x} = {b} , where the square coefficient matrix [A] and the Right-Hand-Side (RHS) vector {b} are known. The unknown solution vector {x} can be efficiently solved by the following step-by-step procedures [1-2]: Reordering phase, Matrix Factorization phase, Forward solution phase, and Backward solution phase. In this research work, a Game-Based Learning (GBL) approach has been developed to help engineering students to understand crucial details about matrix reordering and factorization phases. A "chess-like" game has been developed and can be played by either a single player, or two players. Through this "chess-like" open-ended game, the players/learners will not only understand the key concepts involved in reordering algorithms (based on existing algorithms), but also have the opportunities to "discover new algorithms" which are better than existing algorithms. Implementing the proposed "chess-like" game for matrix reordering and factorization phases can be enhanced by FLASH [3] computer environments, where computer simulation with animated human voice, sound effects, visual/graphical/colorful displays of matrix tables, score (or monetary) awards for the best game players, etc. can all be exploited. Preliminary demonstrations of the developed GBL approach can be viewed by anyone who has access to the internet web-site [4]!
Safari, A.; Sharifi, M. A.; Amjadiparvar, B.
2010-05-01
The GRACE mission has substantiated the low-low satellite-to-satellite tracking (LL-SST) concept. The LL-SST configuration can be combined with the previously realized high-low SST concept in the CHAMP mission to provide a much higher accuracy. The line of sight (LOS) acceleration difference between the GRACE satellite pair is the mostly used observable for mapping the global gravity field of the Earth in terms of spherical harmonic coefficients. In this paper, mathematical formulae for LOS acceleration difference observations have been derived and the corresponding linear system of equations has been set up for spherical harmonic up to degree and order 120. The total number of unknowns is 14641. Such a linear equation system can be solved with iterative solvers or direct solvers. However, the runtime of direct methods or that of iterative solvers without a suitable preconditioner increases tremendously. This is the reason why we need a more sophisticated method to solve the linear system of problems with a large number of unknowns. Multiplicative variant of the Schwarz alternating algorithm is a domain decomposition method, which allows it to split the normal matrix of the system into several smaller overlaped submatrices. In each iteration step the multiplicative variant of the Schwarz alternating algorithm solves linear systems with the matrices obtained from the splitting successively. It reduces both runtime and memory requirements drastically. In this paper we propose the Multiplicative Schwarz Alternating Algorithm (MSAA) for solving the large linear system of gravity field recovery. The proposed algorithm has been tested on the International Association of Geodesy (IAG)-simulated data of the GRACE mission. The achieved results indicate the validity and efficiency of the proposed algorithm in solving the linear system of equations from accuracy and runtime points of view. Keywords: Gravity field recovery, Multiplicative Schwarz Alternating Algorithm, Low
Carpentieri, Bruno; Jing, Yan-Fei; Huang, Ting-Zhu; Pi, Wei-Chao; Sheng, Xin-Qing
We report on experiments with a novel family of Krylov subspace methods for solving dense, complex, non-Hermitian systems of linear equations arising from the Galerkin discretization of surface integral equation models in Electromagnetics. By some experiments on realistic radar-cross-section
Students’ difficulties in solving linear equation problems
Wati, S.; Fitriana, L.; Mardiyana
2018-03-01
A linear equation is an algebra material that exists in junior high school to university. It is a very important material for students in order to learn more advanced mathematics topics. Therefore, linear equation material is essential to be mastered. However, the result of 2016 national examination in Indonesia showed that students’ achievement in solving linear equation problem was low. This fact became a background to investigate students’ difficulties in solving linear equation problems. This study used qualitative descriptive method. An individual written test on linear equation tasks was administered, followed by interviews. Twenty-one sample students of grade VIII of SMPIT Insan Kamil Karanganyar did the written test, and 6 of them were interviewed afterward. The result showed that students with high mathematics achievement donot have difficulties, students with medium mathematics achievement have factual difficulties, and students with low mathematics achievement have factual, conceptual, operational, and principle difficulties. Based on the result there is a need of meaningfulness teaching strategy to help students to overcome difficulties in solving linear equation problems.
Numerical solution of large sparse linear systems
International Nuclear Information System (INIS)
Meurant, Gerard; Golub, Gene.
1982-02-01
This note is based on one of the lectures given at the 1980 CEA-EDF-INRIA Numerical Analysis Summer School whose aim is the study of large sparse linear systems. The main topics are solving least squares problems by orthogonal transformation, fast Poisson solvers and solution of sparse linear system by iterative methods with a special emphasis on preconditioned conjuguate gradient method [fr
Directory of Open Access Journals (Sweden)
Yi-hua Zhong
2013-01-01
Full Text Available Recently, various methods have been developed for solving linear programming problems with fuzzy number, such as simplex method and dual simplex method. But their computational complexities are exponential, which is not satisfactory for solving large-scale fuzzy linear programming problems, especially in the engineering field. A new method which can solve large-scale fuzzy number linear programming problems is presented in this paper, which is named a revised interior point method. Its idea is similar to that of interior point method used for solving linear programming problems in crisp environment before, but its feasible direction and step size are chosen by using trapezoidal fuzzy numbers, linear ranking function, fuzzy vector, and their operations, and its end condition is involved in linear ranking function. Their correctness and rationality are proved. Moreover, choice of the initial interior point and some factors influencing the results of this method are also discussed and analyzed. The result of algorithm analysis and example study that shows proper safety factor parameter, accuracy parameter, and initial interior point of this method may reduce iterations and they can be selected easily according to the actual needs. Finally, the method proposed in this paper is an alternative method for solving fuzzy number linear programming problems.
Solving fault diagnosis problems linear synthesis techniques
Varga, Andreas
2017-01-01
This book addresses fault detection and isolation topics from a computational perspective. Unlike most existing literature, it bridges the gap between the existing well-developed theoretical results and the realm of reliable computational synthesis procedures. The model-based approach to fault detection and diagnosis has been the subject of ongoing research for the past few decades. While the theoretical aspects of fault diagnosis on the basis of linear models are well understood, most of the computational methods proposed for the synthesis of fault detection and isolation filters are not satisfactory from a numerical standpoint. Several features make this book unique in the fault detection literature: Solution of standard synthesis problems in the most general setting, for both continuous- and discrete-time systems, regardless of whether they are proper or not; consequently, the proposed synthesis procedures can solve a specific problem whenever a solution exists Emphasis on the best numerical algorithms to ...
An approach for solving linear fractional programming problems ...
African Journals Online (AJOL)
The paper presents a new approach for solving a fractional linear programming problem in which the objective function is a linear fractional function, while the constraint functions are in the form of linear inequalities. The approach adopted is based mainly upon solving the problem algebraically using the concept of duality ...
A logic circuit for solving linear function by digital method
International Nuclear Information System (INIS)
Ma Yonghe
1986-01-01
A mathematical method for determining the linear relation of physical quantity with rediation intensity is described. A logic circuit has been designed for solving linear function by digital method. Some applications and the circuit function are discussed
An Approach for Solving Linear Fractional Programming Problems
Andrew Oyakhobo Odior
2012-01-01
Linear fractional programming problems are useful tools in production planning, financial and corporate planning, health care and hospital planning and as such have attracted considerable research interest. The paper presents a new approach for solving a fractional linear programming problem in which the objective function is a linear fractional function, while the constraint functions are in the form of linear inequalities. The approach adopted is based mainly upon solving the problem algebr...
New approach to solve symmetric fully fuzzy linear systems
Indian Academy of Sciences (India)
concepts of fuzzy set theory and then define a fully fuzzy linear system of equations. .... To represent the above problem as fully fuzzy linear system, we represent x .... Fully fuzzy linear systems can be solved by Linear programming approach, ...
Fuzzy linear programming approach for solving transportation
Indian Academy of Sciences (India)
Transportation problem (TP) is an important network structured linear programming problem that arises in several contexts and has deservedly received a great deal of attention in the literature. The central concept in this problem is to find the least total transportation cost of a commodity in order to satisfy demands at ...
Convergence of hybrid methods for solving non-linear partial ...
African Journals Online (AJOL)
This paper is concerned with the numerical solution and convergence analysis of non-linear partial differential equations using a hybrid method. The solution technique involves discretizing the non-linear system of PDE to obtain a corresponding non-linear system of algebraic difference equations to be solved at each time ...
Students' errors in solving linear equation word problems: Case ...
African Journals Online (AJOL)
The study examined errors students make in solving linear equation word problems with a view to expose the nature of these errors and to make suggestions for classroom teaching. A diagnostic test comprising 10 linear equation word problems, was administered to a sample (n=130) of senior high school first year Home ...
Perspectives on large linear colliders
International Nuclear Information System (INIS)
Richter, B.
1987-11-01
Three main items in the design of large linear colliders are presented. The first is the interrelation of energy and luminosity requirements. These two items impose severe constraints on the accelerator builder who must design a machine to meet the needs of experimentl high energy physics rather than designing a machine for its own sake. An introduction is also given for linear collider design, concentrating on what goes on at the collision point, for still another constraint comes here from the beam-beam interaction which further restricts the choices available to the accelerator builder. The author also gives his impressions of the state of the technology available for building these kinds of machines within the next decade. The paper concludes with a brief recommendation for how we can all get on with the work faster, and hope to realize these machines sooner by working together. 10 refs., 9 figs
A Proposed Method for Solving Fuzzy System of Linear Equations
Directory of Open Access Journals (Sweden)
Reza Kargar
2014-01-01
Full Text Available This paper proposes a new method for solving fuzzy system of linear equations with crisp coefficients matrix and fuzzy or interval right hand side. Some conditions for the existence of a fuzzy or interval solution of m×n linear system are derived and also a practical algorithm is introduced in detail. The method is based on linear programming problem. Finally the applicability of the proposed method is illustrated by some numerical examples.
Experimental quantum computing to solve systems of linear equations.
Cai, X-D; Weedbrook, C; Su, Z-E; Chen, M-C; Gu, Mile; Zhu, M-J; Li, Li; Liu, Nai-Le; Lu, Chao-Yang; Pan, Jian-Wei
2013-06-07
Solving linear systems of equations is ubiquitous in all areas of science and engineering. With rapidly growing data sets, such a task can be intractable for classical computers, as the best known classical algorithms require a time proportional to the number of variables N. A recently proposed quantum algorithm shows that quantum computers could solve linear systems in a time scale of order log(N), giving an exponential speedup over classical computers. Here we realize the simplest instance of this algorithm, solving 2×2 linear equations for various input vectors on a quantum computer. We use four quantum bits and four controlled logic gates to implement every subroutine required, demonstrating the working principle of this algorithm.
Approximate Method for Solving the Linear Fuzzy Delay Differential Equations
Directory of Open Access Journals (Sweden)
S. Narayanamoorthy
2015-01-01
Full Text Available We propose an algorithm of the approximate method to solve linear fuzzy delay differential equations using Adomian decomposition method. The detailed algorithm of the approach is provided. The approximate solution is compared with the exact solution to confirm the validity and efficiency of the method to handle linear fuzzy delay differential equation. To show this proper features of this proposed method, numerical example is illustrated.
Solving Fully Fuzzy Linear System of Equations in General Form
Directory of Open Access Journals (Sweden)
A. Yousefzadeh
2012-06-01
Full Text Available In this work, we propose an approach for computing the positive solution of a fully fuzzy linear system where the coefficient matrix is a fuzzy $nimes n$ matrix. To do this, we use arithmetic operations on fuzzy numbers that introduced by Kaffman in and convert the fully fuzzy linear system into two $nimes n$ and $2nimes 2n$ crisp linear systems. If the solutions of these linear systems don't satisfy in positive fuzzy solution condition, we introduce the constrained least squares problem to obtain optimal fuzzy vector solution by applying the ranking function in given fully fuzzy linear system. Using our proposed method, the fully fuzzy linear system of equations always has a solution. Finally, we illustrate the efficiency of proposed method by solving some numerical examples.
New approach to solve symmetric fully fuzzy linear systems
Indian Academy of Sciences (India)
In this paper, we present a method to solve fully fuzzy linear systems with symmetric coefﬁcient matrix. The symmetric coefﬁcient matrix is decomposed into two systems of equations by using Cholesky method and then a solution can be obtained. Numerical examples are given to illustrate our method.
Insights into the School Mathematics Tradition from Solving Linear Equations
Buchbinder, Orly; Chazan, Daniel; Fleming, Elizabeth
2015-01-01
In this article, we explore how the solving of linear equations is represented in English-language algebra text books from the early nineteenth century when schooling was becoming institutionalized, and then survey contemporary teachers. In the text books, we identify the increasing presence of a prescribed order of steps (a canonical method) for…
Students' errors in solving linear equation word problems: Case ...
African Journals Online (AJOL)
kofi.mereku
Development in most areas of life is based on effective knowledge of science and ... Problem solving, as used in mathematics education literature, refers ... word problems, on the other hand, are those linear equation tasks or ... taught LEWPs in the junior high school, many of them reach the senior high school without a.
Galerkin projection methods for solving multiple related linear systems
Energy Technology Data Exchange (ETDEWEB)
Chan, T.F.; Ng, M.; Wan, W.L.
1996-12-31
We consider using Galerkin projection methods for solving multiple related linear systems A{sup (i)}x{sup (i)} = b{sup (i)} for 1 {le} i {le} s, where A{sup (i)} and b{sup (i)} are different in general. We start with the special case where A{sup (i)} = A and A is symmetric positive definite. The method generates a Krylov subspace from a set of direction vectors obtained by solving one of the systems, called the seed system, by the CG method and then projects the residuals of other systems orthogonally onto the generated Krylov subspace to get the approximate solutions. The whole process is repeated with another unsolved system as a seed until all the systems are solved. We observe in practice a super-convergence behaviour of the CG process of the seed system when compared with the usual CG process. We also observe that only a small number of restarts is required to solve all the systems if the right-hand sides are close to each other. These two features together make the method particularly effective. In this talk, we give theoretical proof to justify these observations. Furthermore, we combine the advantages of this method and the block CG method and propose a block extension of this single seed method. The above procedure can actually be modified for solving multiple linear systems A{sup (i)}x{sup (i)} = b{sup (i)}, where A{sup (i)} are now different. We can also extend the previous analytical results to this more general case. Applications of this method to multiple related linear systems arising from image restoration and recursive least squares computations are considered as examples.
Comments on new iterative methods for solving linear systems
Directory of Open Access Journals (Sweden)
Wang Ke
2017-06-01
Full Text Available Some new iterative methods were presented by Du, Zheng and Wang for solving linear systems in [3], where it is shown that the new methods, comparing to the classical Jacobi or Gauss-Seidel method, can be applied to more systems and have faster convergence. This note shows that their methods are suitable for more matrices than positive matrices which the authors suggested through further analysis and numerical examples.
CHEBYSHEV ACCELERATION TECHNIQUE FOR SOLVING FUZZY LINEAR SYSTEM
Directory of Open Access Journals (Sweden)
S.H. Nasseri
2011-07-01
Full Text Available In this paper, Chebyshev acceleration technique is used to solve the fuzzy linear system (FLS. This method is discussed in details and followed by summary of some other acceleration techniques. Moreover, we show that in some situations that the methods such as Jacobi, Gauss-Sidel, SOR and conjugate gradient is divergent, our proposed method is applicable and the acquired results are illustrated by some numerical examples.
CHEBYSHEV ACCELERATION TECHNIQUE FOR SOLVING FUZZY LINEAR SYSTEM
Directory of Open Access Journals (Sweden)
S.H. Nasseri
2009-10-01
Full Text Available In this paper, Chebyshev acceleration technique is used to solve the fuzzy linear system (FLS. This method is discussed in details and followed by summary of some other acceleration techniques. Moreover, we show that in some situations that the methods such as Jacobi, Gauss-Sidel, SOR and conjugate gradient is divergent, our proposed method is applicable and the acquired results are illustrated by some numerical examples.
AZTEC: A parallel iterative package for the solving linear systems
Energy Technology Data Exchange (ETDEWEB)
Hutchinson, S.A.; Shadid, J.N.; Tuminaro, R.S. [Sandia National Labs., Albuquerque, NM (United States)
1996-12-31
We describe a parallel linear system package, AZTEC. The package incorporates a number of parallel iterative methods (e.g. GMRES, biCGSTAB, CGS, TFQMR) and preconditioners (e.g. Jacobi, Gauss-Seidel, polynomial, domain decomposition with LU or ILU within subdomains). Additionally, AZTEC allows for the reuse of previous preconditioning factorizations within Newton schemes for nonlinear methods. Currently, a number of different users are using this package to solve a variety of PDE applications.
Solving linear inequalities in a least squares sense
Energy Technology Data Exchange (ETDEWEB)
Bramley, R.; Winnicka, B. [Indiana Univ., Bloomington, IN (United States)
1994-12-31
Let A {element_of} {Re}{sup mxn} be an arbitrary real matrix, and let b {element_of} {Re}{sup m} a given vector. A familiar problem in computational linear algebra is to solve the system Ax = b in a least squares sense; that is, to find an x* minimizing {parallel}Ax {minus} b{parallel}, where {parallel} {center_dot} {parallel} refers to the vector two-norm. Such an x* solves the normal equations A{sup T}(Ax {minus} b) = 0, and the optimal residual r* = b {minus} Ax* is unique (although x* need not be). The least squares problem is usually interpreted as corresponding to multiple observations, represented by the rows of A and b, on a vector of data x. The observations may be inconsistent, and in this case a solution is sought that minimizes the norm of the residuals. A less familiar problem to numerical linear algebraists is the solution of systems of linear inequalities Ax {le} b in a least squares sense, but the motivation is similar: if a set of observations places upper or lower bounds on linear combinations of variables, the authors want to find x* minimizing {parallel} (Ax {minus} b){sub +} {parallel}, where the i{sup th} component of the vector v{sub +} is the maximum of zero and the i{sup th} component of v.
Perspectives on large Linear Colliders
International Nuclear Information System (INIS)
Richter, B.
1987-01-01
The accelerator community now generally agrees that the Linear Collider is the most cost-effective technology for reaching much higher energies in the center-of-mass than can be attained in the largest of the e + e - storage rings, LEP. Indeed, even as the first linear collider, the SLC at SLAC, is getting ready to begin operations groups, at SLAC, Novosibirsk, CERN and KEK are doing R and D and conceptual design studies on a next generation machine in the 1 TeV energy region. In this perspectives talk I do not want to restrict my comments to any particular design, and so I will talk about a high-energy machine as the NLC, which is shorthand for the Next Linear Collider, and taken to mean a machine with a center-of-mass energy someplace in the 0.5 to 2 TeV energy range with sufficient luminosity to carry out a meaningful experimental program. I want to discuss three main items with you. The first is the interrelation of energy and luminosity requirements. These two items impose severe constraints on the accelerator builder. Next, I will give an introduction to linear collider design, concentrating on what goes on at the collision point, for still another constraint comes here from the beam-beam interaction which further restricts the choices available to the accelerator builder.Then, I want to give my impressions of the state of the technology available for building these kinds of machines within the next decade
SLAP, Large Sparse Linear System Solution Package
International Nuclear Information System (INIS)
Greenbaum, A.
1987-01-01
1 - Description of program or function: SLAP is a set of routines for solving large sparse systems of linear equations. One need not store the entire matrix - only the nonzero elements and their row and column numbers. Any nonzero structure is acceptable, so the linear system solver need not be modified when the structure of the matrix changes. Auxiliary storage space is acquired and released within the routines themselves by use of the LRLTRAN POINTER statement. 2 - Method of solution: SLAP contains one direct solver, a band matrix factorization and solution routine, BAND, and several interactive solvers. The iterative routines are as follows: JACOBI, Jacobi iteration; GS, Gauss-Seidel Iteration; ILUIR, incomplete LU decomposition with iterative refinement; DSCG and ICCG, diagonal scaling and incomplete Cholesky decomposition with conjugate gradient iteration (for symmetric positive definite matrices only); DSCGN and ILUGGN, diagonal scaling and incomplete LU decomposition with conjugate gradient interaction on the normal equations; DSBCG and ILUBCG, diagonal scaling and incomplete LU decomposition with bi-conjugate gradient iteration; and DSOMN and ILUOMN, diagonal scaling and incomplete LU decomposition with ORTHOMIN iteration
Mathematics Literacy of Secondary Students in Solving Simultanenous Linear Equations
Sitompul, R. S. I.; Budayasa, I. K.; Masriyah
2018-01-01
This study examines the profile of secondary students’ mathematical literacy in solving simultanenous linear equations problems in terms of cognitive style of visualizer and verbalizer. This research is a descriptive research with qualitative approach. The subjects in this research consist of one student with cognitive style of visualizer and one student with cognitive style of verbalizer. The main instrument in this research is the researcher herself and supporting instruments are cognitive style tests, mathematics skills tests, problem-solving tests and interview guidelines. Research was begun by determining the cognitive style test and mathematics skill test. The subjects chosen were given problem-solving test about simultaneous linear equations and continued with interview. To ensure the validity of the data, the researcher conducted data triangulation; the steps of data reduction, data presentation, data interpretation, and conclusion drawing. The results show that there is a similarity of visualizer and verbalizer-cognitive style in identifying and understanding the mathematical structure in the process of formulating. There are differences in how to represent problems in the process of implementing, there are differences in designing strategies and in the process of interpreting, and there are differences in explaining the logical reasons.
A recurrent neural network for solving bilevel linear programming problem.
He, Xing; Li, Chuandong; Huang, Tingwen; Li, Chaojie; Huang, Junjian
2014-04-01
In this brief, based on the method of penalty functions, a recurrent neural network (NN) modeled by means of a differential inclusion is proposed for solving the bilevel linear programming problem (BLPP). Compared with the existing NNs for BLPP, the model has the least number of state variables and simple structure. Using nonsmooth analysis, the theory of differential inclusions, and Lyapunov-like method, the equilibrium point sequence of the proposed NNs can approximately converge to an optimal solution of BLPP under certain conditions. Finally, the numerical simulations of a supply chain distribution model have shown excellent performance of the proposed recurrent NNs.
New approach to solve fully fuzzy system of linear equations using ...
Indian Academy of Sciences (India)
Known example problems are solved to illustrate the efficacy and ... The concept of fuzzy set and fuzzy number were first introduced by Zadeh .... (iii) Fully fuzzy linear systems can be solved by linear programming approach, Gauss elim-.
Solving linear systems in FLICA-4, thermohydraulic code for 3-D transient computations
International Nuclear Information System (INIS)
Allaire, G.
1995-01-01
FLICA-4 is a computer code, developed at the CEA (France), devoted to steady state and transient thermal-hydraulic analysis of nuclear reactor cores, for small size problems (around 100 mesh cells) as well as for large ones (more than 100000), on, either standard workstations or vector super-computers. As for time implicit codes, the largest time and memory consuming part of FLICA-4 is the routine dedicated to solve the linear system (the size of which is of the order of the number of cells). Therefore, the efficiency of the code is crucially influenced by the optimization of the algorithms used in assembling and solving linear systems: direct methods as the Gauss (or LU) decomposition for moderate size problems, iterative methods as the preconditioned conjugate gradient for large problems. 6 figs., 13 refs
Parallel computation for solving the tridiagonal linear system of equations
International Nuclear Information System (INIS)
Ishiguro, Misako; Harada, Hiroo; Fujii, Minoru; Fujimura, Toichiro; Nakamura, Yasuhiro; Nanba, Katsumi.
1981-09-01
Recently, applications of parallel computation for scientific calculations have increased from the need of the high speed calculation of large scale programs. At the JAERI computing center, an array processor FACOM 230-75 APU has installed to study the applicability of parallel computation for nuclear codes. We made some numerical experiments by using the APU on the methods of solution of tridiagonal linear equation which is an important problem in scientific calculations. Referring to the recent papers with parallel methods, we investigate eight ones. These are Gauss elimination method, Parallel Gauss method, Accelerated parallel Gauss method, Jacobi method, Recursive doubling method, Cyclic reduction method, Chebyshev iteration method, and Conjugate gradient method. The computing time and accuracy were compared among the methods on the basis of the numerical experiments. As the result, it is found that the Cyclic reduction method is best both in computing time and accuracy and the Gauss elimination method is the second one. (author)
Ten-Year-Old Students Solving Linear Equations
Brizuela, Barbara; Schliemann, Analucia
2004-01-01
In this article, the authors seek to re-conceptualize the perspective regarding students' difficulties with algebra. While acknowledging that students "do" have difficulties when learning algebra, they also argue that the generally espoused criteria for algebra as the ability to work with the syntactical rules for solving equations is…
Essential linear algebra with applications a problem-solving approach
Andreescu, Titu
2014-01-01
This textbook provides a rigorous introduction to linear algebra in addition to material suitable for a more advanced course while emphasizing the subject’s interactions with other topics in mathematics such as calculus and geometry. A problem-based approach is used to develop the theoretical foundations of vector spaces, linear equations, matrix algebra, eigenvectors, and orthogonality. Key features include: • a thorough presentation of the main results in linear algebra along with numerous examples to illustrate the theory; • over 500 problems (half with complete solutions) carefully selected for their elegance and theoretical significance; • an interleaved discussion of geometry and linear algebra, giving readers a solid understanding of both topics and the relationship between them. Numerous exercises and well-chosen examples make this text suitable for advanced courses at the junior or senior levels. It can also serve as a source of supplementary problems for a sophomore-level course. ...
A program package for solving linear optimization problems
International Nuclear Information System (INIS)
Horikami, Kunihiko; Fujimura, Toichiro; Nakahara, Yasuaki
1980-09-01
Seven computer programs for the solution of linear, integer and quadratic programming (four programs for linear programming, one for integer programming and two for quadratic programming) have been prepared and tested on FACOM M200 computer, and auxiliary programs have been written to make it easy to use the optimization program package. The characteristics of each program are explained and the detailed input/output descriptions are given in order to let users know how to use them. (author)
Solution methods for large systems of linear equations in BACCHUS
International Nuclear Information System (INIS)
Homann, C.; Dorr, B.
1993-05-01
The computer programme BACCHUS is used to describe steady state and transient thermal-hydraulic behaviour of a coolant in a fuel element with intact geometry in a fast breeder reactor. In such computer programmes generally large systems of linear equations with sparse matrices of coefficients, resulting from discretization of coolant conservation equations, must be solved thousands of times giving rise to large demands of main storage and CPU time. Direct and iterative solution methods of the systems of linear equations, available in BACCHUS, are described, giving theoretical details and experience with their use in the programme. Besides use of a method of lines, a Runge-Kutta-method, for solution of the partial differential equation is outlined. (orig.) [de
Efficient Implementation of the Riccati Recursion for Solving Linear-Quadratic Control Problems
DEFF Research Database (Denmark)
Frison, Gianluca; Jørgensen, John Bagterp
2013-01-01
In both Active-Set (AS) and Interior-Point (IP) algorithms for Model Predictive Control (MPC), sub-problems in the form of linear-quadratic (LQ) control problems need to be solved at each iteration. The solution of these sub-problems is typically the main computational effort at each iteration....... In this paper, we compare a number of solvers for an extended formulation of the LQ control problem: a Riccati recursion based solver can be considered the best choice for the general problem with dense matrices. Furthermore, we present a novel version of the Riccati solver, that makes use of the Cholesky...... factorization of the Pn matrices to reduce the number of flops. When combined with regularization and mixed precision, this algorithm can solve large instances of the LQ control problem up to 3 times faster than the classical Riccati solver....
Novel algorithm of large-scale simultaneous linear equations
International Nuclear Information System (INIS)
Fujiwara, T; Hoshi, T; Yamamoto, S; Sogabe, T; Zhang, S-L
2010-01-01
We review our recently developed methods of solving large-scale simultaneous linear equations and applications to electronic structure calculations both in one-electron theory and many-electron theory. This is the shifted COCG (conjugate orthogonal conjugate gradient) method based on the Krylov subspace, and the most important issue for applications is the shift equation and the seed switching method, which greatly reduce the computational cost. The applications to nano-scale Si crystals and the double orbital extended Hubbard model are presented.
High-order quantum algorithm for solving linear differential equations
International Nuclear Information System (INIS)
Berry, Dominic W
2014-01-01
Linear differential equations are ubiquitous in science and engineering. Quantum computers can simulate quantum systems, which are described by a restricted type of linear differential equations. Here we extend quantum simulation algorithms to general inhomogeneous sparse linear differential equations, which describe many classical physical systems. We examine the use of high-order methods (where the error over a time step is a high power of the size of the time step) to improve the efficiency. These provide scaling close to Δt 2 in the evolution time Δt. As with other algorithms of this type, the solution is encoded in amplitudes of the quantum state, and it is possible to extract global features of the solution. (paper)
Solving Large Clustering Problems with Meta-Heuristic Search
DEFF Research Database (Denmark)
Turkensteen, Marcel; Andersen, Kim Allan; Bang-Jensen, Jørgen
In Clustering Problems, groups of similar subjects are to be retrieved from data sets. In this paper, Clustering Problems with the frequently used Minimum Sum-of-Squares Criterion are solved using meta-heuristic search. Tabu search has proved to be a successful methodology for solving optimization...... problems, but applications to large clustering problems are rare. The simulated annealing heuristic has mainly been applied to relatively small instances. In this paper, we implement tabu search and simulated annealing approaches and compare them to the commonly used k-means approach. We find that the meta-heuristic...
The intelligence of dual simplex method to solve linear fractional fuzzy transportation problem.
Narayanamoorthy, S; Kalyani, S
2015-01-01
An approach is presented to solve a fuzzy transportation problem with linear fractional fuzzy objective function. In this proposed approach the fractional fuzzy transportation problem is decomposed into two linear fuzzy transportation problems. The optimal solution of the two linear fuzzy transportations is solved by dual simplex method and the optimal solution of the fractional fuzzy transportation problem is obtained. The proposed method is explained in detail with an example.
The Intelligence of Dual Simplex Method to Solve Linear Fractional Fuzzy Transportation Problem
Directory of Open Access Journals (Sweden)
S. Narayanamoorthy
2015-01-01
Full Text Available An approach is presented to solve a fuzzy transportation problem with linear fractional fuzzy objective function. In this proposed approach the fractional fuzzy transportation problem is decomposed into two linear fuzzy transportation problems. The optimal solution of the two linear fuzzy transportations is solved by dual simplex method and the optimal solution of the fractional fuzzy transportation problem is obtained. The proposed method is explained in detail with an example.
Linearly Ordered Attribute Grammar Scheduling Using SAT-Solving
Bransen, Jeroen; van Binsbergen, L.Thomas; Claessen, Koen; Dijkstra, Atze
2015-01-01
Many computations over trees can be specified using attribute grammars. Compilers for attribute grammars need to find an evaluation order (or schedule) in order to generate efficient code. For the class of linearly ordered attribute grammars such a schedule can be found statically, but this problem
Preconditioned Iterative Methods for Solving Weighted Linear Least Squares Problems
Czech Academy of Sciences Publication Activity Database
Bru, R.; Marín, J.; Mas, J.; Tůma, Miroslav
2014-01-01
Roč. 36, č. 4 (2014), A2002-A2022 ISSN 1064-8275 Institutional support: RVO:67985807 Keywords : preconditioned iterative methods * incomplete decompositions * approximate inverses * linear least squares Subject RIV: BA - General Mathematics Impact factor: 1.854, year: 2014
Hybrid Method for Solving Inventory Problems with a Linear ...
African Journals Online (AJOL)
Osagiede and Omosigho (2004) proposed a direct search method for identifying the number of replenishment when the demand pattern is linearly increasing. The main computational task in this direct search method was associated with finding the optimal number of replenishments. To accelerate the use of this method, the ...
ADM For Solving Linear Second-Order Fredholm Integro-Differential Equations
Karim, Mohd F.; Mohamad, Mahathir; Saifullah Rusiman, Mohd; Che-Him, Norziha; Roslan, Rozaini; Khalid, Kamil
2018-04-01
In this paper, we apply Adomian Decomposition Method (ADM) as numerically analyse linear second-order Fredholm Integro-differential Equations. The approximate solutions of the problems are calculated by Maple package. Some numerical examples have been considered to illustrate the ADM for solving this equation. The results are compared with the existing exact solution. Thus, the Adomian decomposition method can be the best alternative method for solving linear second-order Fredholm Integro-Differential equation. It converges to the exact solution quickly and in the same time reduces computational work for solving the equation. The result obtained by ADM shows the ability and efficiency for solving these equations.
Directory of Open Access Journals (Sweden)
Tunjo Perić
2017-01-01
Full Text Available This paper presents and analyzes the applicability of three linearization techniques used for solving multi-objective linear fractional programming problems using the goal programming method. The three linearization techniques are: (1 Taylor’s polynomial linearization approximation, (2 the method of variable change, and (3 a modification of the method of variable change proposed in [20]. All three linearization techniques are presented and analyzed in two variants: (a using the optimal value of the objective functions as the decision makers’ aspirations, and (b the decision makers’ aspirations are given by the decision makers. As the criteria for the analysis we use the efficiency of the obtained solutions and the difficulties the analyst comes upon in preparing the linearization models. To analyze the applicability of the linearization techniques incorporated in the linear goal programming method we use an example of a financial structure optimization problem.
A new modified conjugate gradient coefficient for solving system of linear equations
Hajar, N.; ‘Aini, N.; Shapiee, N.; Abidin, Z. Z.; Khadijah, W.; Rivaie, M.; Mamat, M.
2017-09-01
Conjugate gradient (CG) method is an evolution of computational method in solving unconstrained optimization problems. This approach is easy to implement due to its simplicity and has been proven to be effective in solving real-life application. Although this field has received copious amount of attentions in recent years, some of the new approaches of CG algorithm cannot surpass the efficiency of the previous versions. Therefore, in this paper, a new CG coefficient which retains the sufficient descent and global convergence properties of the original CG methods is proposed. This new CG is tested on a set of test functions under exact line search. Its performance is then compared to that of some of the well-known previous CG methods based on number of iterations and CPU time. The results show that the new CG algorithm has the best efficiency amongst all the methods tested. This paper also includes an application of the new CG algorithm for solving large system of linear equations
Directory of Open Access Journals (Sweden)
Mohammad Almousa
2013-01-01
Full Text Available The aim of this study is to present the use of a semi analytical method called the optimal homotopy asymptotic method (OHAM for solving the linear Fredholm integral equations of the first kind. Three examples are discussed to show the ability of the method to solve the linear Fredholm integral equations of the first kind. The results indicated that the method is very effective and simple.
Nahay, John Michael
2008-01-01
We present a new approach to solving polynomial ordinary differential equations by transforming them to linear functional equations and then solving the linear functional equations. We will focus most of our attention upon the first-order Abel differential equation with two nonlinear terms in order to demonstrate in as much detail as possible the computations necessary for a complete solution. We mention in our section on further developments that the basic transformation idea can be generali...
Jamali, R. M. Jalal Uddin; Hashem, M. M. A.; Hasan, M. Mahfuz; Rahman, Md. Bazlar
2013-01-01
Solving a set of simultaneous linear equations is probably the most important topic in numerical methods. For solving linear equations, iterative methods are preferred over the direct methods especially when the coefficient matrix is sparse. The rate of convergence of iteration method is increased by using Successive Relaxation (SR) technique. But SR technique is very much sensitive to relaxation factor, {\\omega}. Recently, hybridization of classical Gauss-Seidel based successive relaxation t...
A goal programming procedure for solving fuzzy multiobjective fractional linear programming problems
Directory of Open Access Journals (Sweden)
Tunjo Perić
2014-12-01
Full Text Available This paper presents a modification of Pal, Moitra and Maulik's goal programming procedure for fuzzy multiobjective linear fractional programming problem solving. The proposed modification of the method allows simpler solving of economic multiple objective fractional linear programming (MOFLP problems, enabling the obtained solutions to express the preferences of the decision maker defined by the objective function weights. The proposed method is tested on the production planning example.
M. ZANGIABADI; H. R. MALEKI
2007-01-01
In the real-world optimization problems, coefficients of the objective function are not known precisely and can be interpreted as fuzzy numbers. In this paper we define the concepts of optimality for linear programming problems with fuzzy parameters based on those for multiobjective linear programming problems. Then by using the concept of comparison of fuzzy numbers, we transform a linear programming problem with fuzzy parameters to a multiobjective linear programming problem. To this end, w...
Ergul, Ozgur
2014-01-01
The Multilevel Fast Multipole Algorithm (MLFMA) for Solving Large-Scale Computational Electromagnetic Problems provides a detailed and instructional overview of implementing MLFMA. The book: Presents a comprehensive treatment of the MLFMA algorithm, including basic linear algebra concepts, recent developments on the parallel computation, and a number of application examplesCovers solutions of electromagnetic problems involving dielectric objects and perfectly-conducting objectsDiscusses applications including scattering from airborne targets, scattering from red
Directory of Open Access Journals (Sweden)
Mihai-Victor PRICOP
2010-09-01
Full Text Available The present paper introduces a numerical approach of static linear elasticity equations for anisotropic materials. The domain and boundary conditions are simple, to enhance an easy implementation of the finite difference scheme. SOR and gradient are used to solve the resulting linear system. The simplicity of the geometry is also useful for MPI parallelization of the code.
EZLP: An Interactive Computer Program for Solving Linear Programming Problems. Final Report.
Jarvis, John J.; And Others
Designed for student use in solving linear programming problems, the interactive computer program described (EZLP) permits the student to input the linear programming model in exactly the same manner in which it would be written on paper. This report includes a brief review of the development of EZLP; narrative descriptions of program features,…
New approach to solve fully fuzzy system of linear equations using ...
Indian Academy of Sciences (India)
This paper proposes two new methods to solve fully fuzzy system of linear equations. The fuzzy system has been converted to a crisp system of linear equations by using single and double parametric form of fuzzy numbers to obtain the non-negative solution. Double parametric form of fuzzy numbers is defined and applied ...
Solving non-linear Horn clauses using a linear Horn clause solver
DEFF Research Database (Denmark)
Kafle, Bishoksan; Gallagher, John Patrick; Ganty, Pierre
2016-01-01
In this paper we show that checking satisfiability of a set of non-linear Horn clauses (also called a non-linear Horn clause program) can be achieved using a solver for linear Horn clauses. We achieve this by interleaving a program transformation with a satisfiability checker for linear Horn...... clauses (also called a solver for linear Horn clauses). The program transformation is based on the notion of tree dimension, which we apply to a set of non-linear clauses, yielding a set whose derivation trees have bounded dimension. Such a set of clauses can be linearised. The main algorithm...... dimension. We constructed a prototype implementation of this approach and performed some experiments on a set of verification problems, which shows some promise....
Chosen interval methods for solving linear interval systems with special type of matrix
Szyszka, Barbara
2013-10-01
The paper is devoted to chosen direct interval methods for solving linear interval systems with special type of matrix. This kind of matrix: band matrix with a parameter, from finite difference problem is obtained. Such linear systems occur while solving one dimensional wave equation (Partial Differential Equations of hyperbolic type) by using the central difference interval method of the second order. Interval methods are constructed so as the errors of method are enclosed in obtained results, therefore presented linear interval systems contain elements that determining the errors of difference method. The chosen direct algorithms have been applied for solving linear systems because they have no errors of method. All calculations were performed in floating-point interval arithmetic.
Numerical method for solving linear Fredholm fuzzy integral equations of the second kind
Energy Technology Data Exchange (ETDEWEB)
Abbasbandy, S. [Department of Mathematics, Imam Khomeini International University, P.O. Box 288, Ghazvin 34194 (Iran, Islamic Republic of)]. E-mail: saeid@abbasbandy.com; Babolian, E. [Faculty of Mathematical Sciences and Computer Engineering, Teacher Training University, Tehran 15618 (Iran, Islamic Republic of); Alavi, M. [Department of Mathematics, Arak Branch, Islamic Azad University, Arak 38135 (Iran, Islamic Republic of)
2007-01-15
In this paper we use parametric form of fuzzy number and convert a linear fuzzy Fredholm integral equation to two linear system of integral equation of the second kind in crisp case. We can use one of the numerical method such as Nystrom and find the approximation solution of the system and hence obtain an approximation for fuzzy solution of the linear fuzzy Fredholm integral equations of the second kind. The proposed method is illustrated by solving some numerical examples.
Solving large scale structure in ten easy steps with COLA
Energy Technology Data Exchange (ETDEWEB)
Tassev, Svetlin [Department of Astrophysical Sciences, Princeton University, 4 Ivy Lane, Princeton, NJ 08544 (United States); Zaldarriaga, Matias [School of Natural Sciences, Institute for Advanced Study, Olden Lane, Princeton, NJ 08540 (United States); Eisenstein, Daniel J., E-mail: stassev@cfa.harvard.edu, E-mail: matiasz@ias.edu, E-mail: deisenstein@cfa.harvard.edu [Center for Astrophysics, Harvard University, 60 Garden Street, Cambridge, MA 02138 (United States)
2013-06-01
We present the COmoving Lagrangian Acceleration (COLA) method: an N-body method for solving for Large Scale Structure (LSS) in a frame that is comoving with observers following trajectories calculated in Lagrangian Perturbation Theory (LPT). Unlike standard N-body methods, the COLA method can straightforwardly trade accuracy at small-scales in order to gain computational speed without sacrificing accuracy at large scales. This is especially useful for cheaply generating large ensembles of accurate mock halo catalogs required to study galaxy clustering and weak lensing, as those catalogs are essential for performing detailed error analysis for ongoing and future surveys of LSS. As an illustration, we ran a COLA-based N-body code on a box of size 100 Mpc/h with particles of mass ≈ 5 × 10{sup 9}M{sub s}un/h. Running the code with only 10 timesteps was sufficient to obtain an accurate description of halo statistics down to halo masses of at least 10{sup 11}M{sub s}un/h. This is only at a modest speed penalty when compared to mocks obtained with LPT. A standard detailed N-body run is orders of magnitude slower than our COLA-based code. The speed-up we obtain with COLA is due to the fact that we calculate the large-scale dynamics exactly using LPT, while letting the N-body code solve for the small scales, without requiring it to capture exactly the internal dynamics of halos. Achieving a similar level of accuracy in halo statistics without the COLA method requires at least 3 times more timesteps than when COLA is employed.
International Nuclear Information System (INIS)
Ravi Kanth, A.S.V.; Aruna, K.
2009-01-01
In this paper, we propose a reliable algorithm to develop exact and approximate solutions for the linear and nonlinear Schroedinger equations. The approach rest mainly on two-dimensional differential transform method which is one of the approximate methods. The method can easily be applied to many linear and nonlinear problems and is capable of reducing the size of computational work. Exact solutions can also be achieved by the known forms of the series solutions. Several illustrative examples are given to demonstrate the effectiveness of the present method.
A new neural network model for solving random interval linear programming problems.
Arjmandzadeh, Ziba; Safi, Mohammadreza; Nazemi, Alireza
2017-05-01
This paper presents a neural network model for solving random interval linear programming problems. The original problem involving random interval variable coefficients is first transformed into an equivalent convex second order cone programming problem. A neural network model is then constructed for solving the obtained convex second order cone problem. Employing Lyapunov function approach, it is also shown that the proposed neural network model is stable in the sense of Lyapunov and it is globally convergent to an exact satisfactory solution of the original problem. Several illustrative examples are solved in support of this technique. Copyright © 2017 Elsevier Ltd. All rights reserved.
Method for solving fully fuzzy linear programming problems using deviation degree measure
Institute of Scientific and Technical Information of China (English)
Haifang Cheng; Weilai Huang; Jianhu Cai
2013-01-01
A new ful y fuzzy linear programming (FFLP) prob-lem with fuzzy equality constraints is discussed. Using deviation degree measures, the FFLP problem is transformed into a crispδ-parametric linear programming (LP) problem. Giving the value of deviation degree in each constraint, the δ-fuzzy optimal so-lution of the FFLP problem can be obtained by solving this LP problem. An algorithm is also proposed to find a balance-fuzzy optimal solution between two goals in conflict: to improve the va-lues of the objective function and to decrease the values of the deviation degrees. A numerical example is solved to il ustrate the proposed method.
On a new iterative method for solving linear systems and comparison results
Jing, Yan-Fei; Huang, Ting-Zhu
2008-10-01
In Ujevic [A new iterative method for solving linear systems, Appl. Math. Comput. 179 (2006) 725-730], the author obtained a new iterative method for solving linear systems, which can be considered as a modification of the Gauss-Seidel method. In this paper, we show that this is a special case from a point of view of projection techniques. And a different approach is established, which is both theoretically and numerically proven to be better than (at least the same as) Ujevic's. As the presented numerical examples show, in most cases, the convergence rate is more than one and a half that of Ujevic.
Guo, Sangang
2017-09-01
There are two stages in solving security-constrained unit commitment problems (SCUC) within Lagrangian framework: one is to obtain feasible units’ states (UC), the other is power economic dispatch (ED) for each unit. The accurate solution of ED is more important for enhancing the efficiency of the solution to SCUC for the fixed feasible units’ statues. Two novel methods named after Convex Combinatorial Coefficient Method and Power Increment Method respectively based on linear programming problem for solving ED are proposed by the piecewise linear approximation to the nonlinear convex fuel cost functions. Numerical testing results show that the methods are effective and efficient.
Iterative solution of large linear systems
Young, David Matheson
1971-01-01
This self-contained treatment offers a systematic development of the theory of iterative methods. Its focal point resides in an analysis of the convergence properties of the successive overrelaxation (SOR) method, as applied to a linear system with a consistently ordered matrix. The text explores the convergence properties of the SOR method and related techniques in terms of the spectral radii of the associated matrices as well as in terms of certain matrix norms. Contents include a review of matrix theory and general properties of iterative methods; SOR method and stationary modified SOR meth
Radiation protection in large linear accelerators
International Nuclear Information System (INIS)
Oliva, Jose de Jesus Rivero
2013-01-01
The electron linear accelerators can be used in industrial applications that require powerful sources of ionizing radiation. They have the important characteristic of not representing a radiation hazard when the accelerators remain electrically disconnected. With the plant in operation, a high reliability defense in depth reduces the risk of radiological accidents to extremely small levels. It is practically impossible that a person could enter into the radiation bunker with the accelerators connected. Aceletron Irradiacao Industrial, located in Rio de Janeiro, offers services of irradiation by means of two powerful electron linear accelerators, with 15 kW power and 10 MeV electron energy. Despite the high level of existing radiation safety, a simplified risk study is underway to identify possible sequences of radiological accidents. The study is based on the combined application of the event and fault trees techniques. Preliminary results confirm that there is a very small risk of entering into the irradiation bunker with the accelerators in operation, but the risk of an operator entering into the bunker during a process interruption and remaining there without notice after the accelerators were restarted may be considerably larger. Based on these results the Company is considering alternatives to reduce the likelihood of human error of this type that could lead to a radiological accident. The paper describes the defense in depth of the irradiation process in Aceletron Irradiacao Industrial, as well as the models and preliminary results of the ongoing risk analysis, including the additional safety measures which are being evaluated. (author)
Directory of Open Access Journals (Sweden)
Ai-Min Yang
2014-01-01
Full Text Available The local fractional Laplace variational iteration method was applied to solve the linear local fractional partial differential equations. The local fractional Laplace variational iteration method is coupled by the local fractional variational iteration method and Laplace transform. The nondifferentiable approximate solutions are obtained and their graphs are also shown.
The H-N method for solving linear transport equation: theory and application
International Nuclear Information System (INIS)
Kaskas, A.; Gulecyuz, M.C.; Tezcan, C.
2002-01-01
The system of singular integral equation which is obtained from the integro-differential form of the linear transport equation as a result of Placzec lemma is solved. Application are given using the exit distributions and the infinite medium Green's function. The same theoretical results are also obtained with the use of the singular eigenfunction of the method of elementary solutions
Iterative algorithms for large sparse linear systems on parallel computers
Adams, L. M.
1982-01-01
Algorithms for assembling in parallel the sparse system of linear equations that result from finite difference or finite element discretizations of elliptic partial differential equations, such as those that arise in structural engineering are developed. Parallel linear stationary iterative algorithms and parallel preconditioned conjugate gradient algorithms are developed for solving these systems. In addition, a model for comparing parallel algorithms on array architectures is developed and results of this model for the algorithms are given.
Solving Large Scale Nonlinear Eigenvalue Problem in Next-Generation Accelerator Design
Energy Technology Data Exchange (ETDEWEB)
Liao, Ben-Shan; Bai, Zhaojun; /UC, Davis; Lee, Lie-Quan; Ko, Kwok; /SLAC
2006-09-28
A number of numerical methods, including inverse iteration, method of successive linear problem and nonlinear Arnoldi algorithm, are studied in this paper to solve a large scale nonlinear eigenvalue problem arising from finite element analysis of resonant frequencies and external Q{sub e} values of a waveguide loaded cavity in the next-generation accelerator design. They present a nonlinear Rayleigh-Ritz iterative projection algorithm, NRRIT in short and demonstrate that it is the most promising approach for a model scale cavity design. The NRRIT algorithm is an extension of the nonlinear Arnoldi algorithm due to Voss. Computational challenges of solving such a nonlinear eigenvalue problem for a full scale cavity design are outlined.
Analysis of junior high school students' attempt to solve a linear inequality problem
Taqiyuddin, Muhammad; Sumiaty, Encum; Jupri, Al
2017-08-01
Linear inequality is one of fundamental subjects within junior high school mathematics curricula. Several studies have been conducted to asses students' perform on linear inequality. However, it can hardly be found that linear inequality problems are in the form of "ax + b condition leads to the research questions concerning students' attempt on solving a simple linear inequality problem in this form. In order to do so, the written test was administered to 58 students from two schools in Bandung followed by interviews. The other sources of the data are from teachers' interview and mathematics books used by students. After that, the constant comparative method was used to analyse the data. The result shows that the majority approached the question by doing algebraic operations. Interestingly, most of them did it incorrectly. In contrast, algebraic operations were correctly used by some of them. Moreover, the others performed expected-numbers solution, rewriting the question, translating the inequality into words, and blank answer. Furthermore, we found that there is no one who was conscious of the existence of all-numbers solution. It was found that this condition is reasonably due to how little the learning components concern about why a procedure of solving a linear inequality works and possibilities of linear inequality solution.
Matrix form of Legendre polynomials for solving linear integro-differential equations of high order
Kammuji, M.; Eshkuvatov, Z. K.; Yunus, Arif A. M.
2017-04-01
This paper presents an effective approximate solution of high order of Fredholm-Volterra integro-differential equations (FVIDEs) with boundary condition. Legendre truncated series is used as a basis functions to estimate the unknown function. Matrix operation of Legendre polynomials is used to transform FVIDEs with boundary conditions into matrix equation of Fredholm-Volterra type. Gauss Legendre quadrature formula and collocation method are applied to transfer the matrix equation into system of linear algebraic equations. The latter equation is solved by Gauss elimination method. The accuracy and validity of this method are discussed by solving two numerical examples and comparisons with wavelet and methods.
The fastclime Package for Linear Programming and Large-Scale Precision Matrix Estimation in R.
Pang, Haotian; Liu, Han; Vanderbei, Robert
2014-02-01
We develop an R package fastclime for solving a family of regularized linear programming (LP) problems. Our package efficiently implements the parametric simplex algorithm, which provides a scalable and sophisticated tool for solving large-scale linear programs. As an illustrative example, one use of our LP solver is to implement an important sparse precision matrix estimation method called CLIME (Constrained L 1 Minimization Estimator). Compared with existing packages for this problem such as clime and flare, our package has three advantages: (1) it efficiently calculates the full piecewise-linear regularization path; (2) it provides an accurate dual certificate as stopping criterion; (3) it is completely coded in C and is highly portable. This package is designed to be useful to statisticians and machine learning researchers for solving a wide range of problems.
Shen, Peiping; Zhang, Tongli; Wang, Chunfeng
2017-01-01
This article presents a new approximation algorithm for globally solving a class of generalized fractional programming problems (P) whose objective functions are defined as an appropriate composition of ratios of affine functions. To solve this problem, the algorithm solves an equivalent optimization problem (Q) via an exploration of a suitably defined nonuniform grid. The main work of the algorithm involves checking the feasibility of linear programs associated with the interesting grid points. It is proved that the proposed algorithm is a fully polynomial time approximation scheme as the ratio terms are fixed in the objective function to problem (P), based on the computational complexity result. In contrast to existing results in literature, the algorithm does not require the assumptions on quasi-concavity or low-rank of the objective function to problem (P). Numerical results are given to illustrate the feasibility and effectiveness of the proposed algorithm.
Projective-Dual Method for Solving Systems of Linear Equations with Nonnegative Variables
Ganin, B. V.; Golikov, A. I.; Evtushenko, Yu. G.
2018-02-01
In order to solve an underdetermined system of linear equations with nonnegative variables, the projection of a given point onto its solutions set is sought. The dual of this problem—the problem of unconstrained maximization of a piecewise-quadratic function—is solved by Newton's method. The problem of unconstrained optimization dual of the regularized problem of finding the projection onto the solution set of the system is considered. A connection of duality theory and Newton's method with some known algorithms of projecting onto a standard simplex is shown. On the example of taking into account the specifics of the constraints of the transport linear programming problem, the possibility to increase the efficiency of calculating the generalized Hessian matrix is demonstrated. Some examples of numerical calculations using MATLAB are presented.
Directory of Open Access Journals (Sweden)
Salih Yalcinbas
2016-01-01
Full Text Available In this paper, a new collocation method based on the Fibonacci polynomials is introduced to solve the high-order linear Volterra integro-differential equations under the conditions. Numerical examples are included to demonstrate the applicability and validity of the proposed method and comparisons are made with the existing results. In addition, an error estimation based on the residual functions is presented for this method. The approximate solutions are improved by using this error estimation.
Stress-constrained truss topology optimization problems that can be solved by linear programming
DEFF Research Database (Denmark)
Stolpe, Mathias; Svanberg, Krister
2004-01-01
We consider the problem of simultaneously selecting the material and determining the area of each bar in a truss structure in such a way that the cost of the structure is minimized subject to stress constraints under a single load condition. We show that such problems can be solved by linear...... programming to give the global optimum, and that two different materials are always sufficient in an optimal structure....
Cichocki, A; Unbehauen, R
1994-01-01
In this paper a new class of simplified low-cost analog artificial neural networks with on chip adaptive learning algorithms are proposed for solving linear systems of algebraic equations in real time. The proposed learning algorithms for linear least squares (LS), total least squares (TLS) and data least squares (DLS) problems can be considered as modifications and extensions of well known algorithms: the row-action projection-Kaczmarz algorithm and/or the LMS (Adaline) Widrow-Hoff algorithms. The algorithms can be applied to any problem which can be formulated as a linear regression problem. The correctness and high performance of the proposed neural networks are illustrated by extensive computer simulation results.
Large-scale linear programs in planning and prediction.
2017-06-01
Large-scale linear programs are at the core of many traffic-related optimization problems in both planning and prediction. Moreover, many of these involve significant uncertainty, and hence are modeled using either chance constraints, or robust optim...
A parallel algorithm for solving linear equations arising from one-dimensional network problems
International Nuclear Information System (INIS)
Mesina, G.L.
1991-01-01
One-dimensional (1-D) network problems, such as those arising from 1- D fluid simulations and electrical circuitry, produce systems of sparse linear equations which are nearly tridiagonal and contain a few non-zero entries outside the tridiagonal. Most direct solution techniques for such problems either do not take advantage of the special structure of the matrix or do not fully utilize parallel computer architectures. We describe a new parallel direct linear equation solution algorithm, called TRBR, which is especially designed to take advantage of this structure on MIMD shared memory machines. The new method belongs to a family of methods which split the coefficient matrix into the sum of a tridiagonal matrix T and a matrix comprised of the remaining coefficients R. Efficient tridiagonal methods are used to algebraically simplify the linear system. A smaller auxiliary subsystem is created and solved and its solution is used to calculate the solution of the original system. The newly devised BR method solves the subsystem. The serial and parallel operation counts are given for the new method and related earlier methods. TRBR is shown to have the smallest operation count in this class of direct methods. Numerical results are given. Although the algorithm is designed for one-dimensional networks, it has been applied successfully to three-dimensional problems as well. 20 refs., 2 figs., 4 tabs
International Nuclear Information System (INIS)
Huang Zhenghai; Gu Weizhe
2008-01-01
In this paper, we construct an augmented system of the standard monotone linear complementarity problem (LCP), and establish the relations between the augmented system and the LCP. We present a smoothing-type algorithm for solving the augmented system. The algorithm is shown to be globally convergent without assuming any prior knowledge of feasibility/infeasibility of the problem. In particular, if the LCP has a solution, then the algorithm either generates a maximal complementary solution of the LCP or detects correctly solvability of the LCP, and in the latter case, an existing smoothing-type algorithm can be directly applied to solve the LCP without any additional assumption and it generates a maximal complementary solution of the LCP; and that if the LCP is infeasible, then the algorithm detect correctly infeasibility of the LCP. To the best of our knowledge, such properties have not appeared in the existing literature for smoothing-type algorithms
Scilab software as an alternative low-cost computing in solving the linear equations problem
Agus, Fahrul; Haviluddin
2017-02-01
Numerical computation packages are widely used both in teaching and research. These packages consist of license (proprietary) and open source software (non-proprietary). One of the reasons to use the package is a complexity of mathematics function (i.e., linear problems). Also, number of variables in a linear or non-linear function has been increased. The aim of this paper was to reflect on key aspects related to the method, didactics and creative praxis in the teaching of linear equations in higher education. If implemented, it could be contribute to a better learning in mathematics area (i.e., solving simultaneous linear equations) that essential for future engineers. The focus of this study was to introduce an additional numerical computation package of Scilab as an alternative low-cost computing programming. In this paper, Scilab software was proposed some activities that related to the mathematical models. In this experiment, four numerical methods such as Gaussian Elimination, Gauss-Jordan, Inverse Matrix, and Lower-Upper Decomposition (LU) have been implemented. The results of this study showed that a routine or procedure in numerical methods have been created and explored by using Scilab procedures. Then, the routine of numerical method that could be as a teaching material course has exploited.
Development and adjustment of programs for solving systems of linear equations
International Nuclear Information System (INIS)
Fujimura, Toichiro
1978-03-01
Programs for solving the systems of linear equations have been adjusted and developed in expanding the scientific subroutine library SSL. The principal programs adjusted are based on the congruent method, method of product form of the inverse, orthogonal method, Crout's method for sparse system, and acceleration of iterative methods. The programs developed are based on the escalator method, direct parallel residue method and block tridiagonal method for band system. Described are usage of the programs developed and their future improvement. FORTRAN lists with simple examples in tests of the programs are also given. (auth.)
Parallel Implementation of Riccati Recursion for Solving Linear-Quadratic Control Problems
DEFF Research Database (Denmark)
Frison, Gianluca; Jørgensen, John Bagterp
2013-01-01
In both Active-Set (AS) and Interior-Point (IP) algorithms for Model Predictive Control (MPC), sub-problems in the form of linear-quadratic (LQ) control problems need to be solved at each iteration. The solution of these sub-problems is usually the main computational effort. In this paper...... an alternative version of the Riccati recursion solver for LQ control problems is presented. The performance of both the classical and the alternative version is analyzed from a theoretical as well as a numerical point of view, and the alternative version is found to be approximately 50% faster than...
Directory of Open Access Journals (Sweden)
Animesh Biswas
2016-04-01
Full Text Available This paper deals with fuzzy goal programming approach to solve fuzzy linear bilevel integer programming problems with fuzzy probabilistic constraints following Pareto distribution and Frechet distribution. In the proposed approach a new chance constrained programming methodology is developed from the view point of managing those probabilistic constraints in a hybrid fuzzy environment. A method of defuzzification of fuzzy numbers using ?-cut has been adopted to reduce the problem into a linear bilevel integer programming problem. The individual optimal value of the objective of each DM is found in isolation to construct the fuzzy membership goals. Finally, fuzzy goal programming approach is used to achieve maximum degree of each of the membership goals by minimizing under deviational variables in the decision making environment. To demonstrate the efficiency of the proposed approach, a numerical example is provided.
Solving Large Scale Crew Scheduling Problems in Practice
E.J.W. Abbink (Erwin); L. Albino; T.A.B. Dollevoet (Twan); D. Huisman (Dennis); J. Roussado; R.L. Saldanha
2010-01-01
textabstractThis paper deals with large-scale crew scheduling problems arising at the Dutch railway operator, Netherlands Railways (NS). NS operates about 30,000 trains a week. All these trains need a driver and a certain number of guards. Some labor rules restrict the duties of a certain crew base
An Improved Method for Solving Multiobjective Integer Linear Fractional Programming Problem
Directory of Open Access Journals (Sweden)
Meriem Ait Mehdi
2014-01-01
Full Text Available We describe an improvement of Chergui and Moulaï’s method (2008 that generates the whole efficient set of a multiobjective integer linear fractional program based on the branch and cut concept. The general step of this method consists in optimizing (maximizing without loss of generality one of the fractional objective functions over a subset of the original continuous feasible set; then if necessary, a branching process is carried out until obtaining an integer feasible solution. At this stage, an efficient cut is built from the criteria’s growth directions in order to discard a part of the feasible domain containing only nonefficient solutions. Our contribution concerns firstly the optimization process where a linear program that we define later will be solved at each step rather than a fractional linear program. Secondly, local ideal and nadir points will be used as bounds to prune some branches leading to nonefficient solutions. The computational experiments show that the new method outperforms the old one in all the treated instances.
Solving Large Quadratic|Assignment Problems in Parallel
DEFF Research Database (Denmark)
Clausen, Jens; Perregaard, Michael
1997-01-01
and recalculation of bounds between branchings when used in a parallel Branch-and-Bound algorithm. The algorithm has been implemented on a 16-processor MEIKO Computing Surface with Intel i860 processors. Computational results from the solution of a number of large QAPs, including the classical Nugent 20...... processors, and have hence not been ideally suited for computations essentially involving non-vectorizable computations on integers.In this paper we investigate the combination of one of the best bound functions for a Branch-and-Bound algorithm (the Gilmore-Lawler bound) and various testing, variable binding...
Estimation and Inference for Very Large Linear Mixed Effects Models
Gao, K.; Owen, A. B.
2016-01-01
Linear mixed models with large imbalanced crossed random effects structures pose severe computational problems for maximum likelihood estimation and for Bayesian analysis. The costs can grow as fast as $N^{3/2}$ when there are N observations. Such problems arise in any setting where the underlying factors satisfy a many to many relationship (instead of a nested one) and in electronic commerce applications, the N can be quite large. Methods that do not account for the correlation structure can...
Penalized Estimation in Large-Scale Generalized Linear Array Models
DEFF Research Database (Denmark)
Lund, Adam; Vincent, Martin; Hansen, Niels Richard
2017-01-01
Large-scale generalized linear array models (GLAMs) can be challenging to fit. Computation and storage of its tensor product design matrix can be impossible due to time and memory constraints, and previously considered design matrix free algorithms do not scale well with the dimension...
A composite step conjugate gradients squared algorithm for solving nonsymmetric linear systems
Chan, Tony; Szeto, Tedd
1994-03-01
We propose a new and more stable variant of the CGS method [27] for solving nonsymmetric linear systems. The method is based on squaring the Composite Step BCG method, introduced recently by Bank and Chan [1,2], which itself is a stabilized variant of BCG in that it skips over steps for which the BCG iterate is not defined and causes one kind of breakdown in BCG. By doing this, we obtain a method (Composite Step CGS or CSCGS) which not only handles the breakdowns described above, but does so with the advantages of CGS, namely, no multiplications by the transpose matrix and a faster convergence rate than BCG. Our strategy for deciding whether to skip a step does not involve any machine dependent parameters and is designed to skip near breakdowns as well as produce smoother iterates. Numerical experiments show that the new method does produce improved performance over CGS on practical problems.
Non-linear algorithms solved with the help of the GIBIANE macro-language
International Nuclear Information System (INIS)
Ebersolt, L.; Combescure, A.; Millard, A.; Verpeaux, P.
1987-01-01
Non linear finite element problems are often solved with the help of iteratives procedures. In the finite element program CASTEM 2000, the syntax of the dataset permits the user to derive his own algorithm and tune it to his problem. These basic ideas, simple to imagine, needed a proper frame to be materialized in a general purpose finite element program, and three concepts emerged: Operators, the Gibiane macro-language. In the two first paragraphs, we will detail these concepts, in the third paragraph, we will describe the different possibilities of the program, in the fourth paragraph, we will show, by combining operators in a proper order, how to obtain the desired algorithm. (orig./GL)
DEFF Research Database (Denmark)
Barari, Amin; Ganjavi, B.; Jeloudar, M. Ghanbari
2010-01-01
and fluid mechanics. Design/methodology/approach – Two new but powerful analytical methods, namely, He's VIM and HPM, are introduced to solve some boundary value problems in structural engineering and fluid mechanics. Findings – Analytical solutions often fit under classical perturbation methods. However......, as with other analytical techniques, certain limitations restrict the wide application of perturbation methods, most important of which is the dependence of these methods on the existence of a small parameter in the equation. Disappointingly, the majority of nonlinear problems have no small parameter at all......Purpose – In the last two decades with the rapid development of nonlinear science, there has appeared ever-increasing interest of scientists and engineers in the analytical techniques for nonlinear problems. This paper considers linear and nonlinear systems that are not only regarded as general...
Solving block linear systems with low-rank off-diagonal blocks is easily parallelizable
Energy Technology Data Exchange (ETDEWEB)
Menkov, V. [Indiana Univ., Bloomington, IN (United States)
1996-12-31
An easily and efficiently parallelizable direct method is given for solving a block linear system Bx = y, where B = D + Q is the sum of a non-singular block diagonal matrix D and a matrix Q with low-rank blocks. This implicitly defines a new preconditioning method with an operation count close to the cost of calculating a matrix-vector product Qw for some w, plus at most twice the cost of calculating Qw for some w. When implemented on a parallel machine the processor utilization can be as good as that of those operations. Order estimates are given for the general case, and an implementation is compared to block SSOR preconditioning.
International Nuclear Information System (INIS)
Hernandez-Walls, R; Martín-Atienza, B; Salinas-Matus, M; Castillo, J
2017-01-01
When solving the linear inviscid shallow water equations with variable depth in one dimension using finite differences, a tridiagonal system of equations must be solved. Here we present an approach, which is more efficient than the commonly used numerical method, to solve this tridiagonal system of equations using a recursion formula. We illustrate this approach with an example in which we solve for a rectangular channel to find the resonance modes. Our numerical solution agrees very well with the analytical solution. This new method is easy to use and understand by undergraduate students, so it can be implemented in undergraduate courses such as Numerical Methods, Lineal Algebra or Differential Equations. (paper)
Hernandez-Walls, R.; Martín-Atienza, B.; Salinas-Matus, M.; Castillo, J.
2017-11-01
When solving the linear inviscid shallow water equations with variable depth in one dimension using finite differences, a tridiagonal system of equations must be solved. Here we present an approach, which is more efficient than the commonly used numerical method, to solve this tridiagonal system of equations using a recursion formula. We illustrate this approach with an example in which we solve for a rectangular channel to find the resonance modes. Our numerical solution agrees very well with the analytical solution. This new method is easy to use and understand by undergraduate students, so it can be implemented in undergraduate courses such as Numerical Methods, Lineal Algebra or Differential Equations.
Directory of Open Access Journals (Sweden)
A. Aminataei
2014-05-01
Full Text Available In this paper, a new and ecient approach is applied for numerical approximation of the linear dierential equations with variable coecients based on operational matrices with respect to Hermite polynomials. Explicit formulae which express the Hermite expansioncoecients for the moments of derivatives of any dierentiable function in terms of the original expansion coecients of the function itself are given in the matrix form. The mainimportance of this scheme is that using this approach reduces solving the linear dierentialequations to solve a system of linear algebraic equations, thus greatly simplifying the problem. In addition, two experiments are given to demonstrate the validity and applicability of the method
A novel algebraic procedure for solving non-linear evolution equations of higher order
International Nuclear Information System (INIS)
Huber, Alfred
2007-01-01
We report here a systematic approach that can easily be used for solving non-linear partial differential equations (nPDE), especially of higher order. We restrict the analysis to the so called evolution equations describing any wave propagation. The proposed new algebraic approach leads us to traveling wave solutions and moreover, new class of solution can be obtained. The crucial step of our method is the basic assumption that the solutions satisfy an ordinary differential equation (ODE) of first order that can be easily integrated. The validity and reliability of the method is tested by its application to some non-linear evolution equations. The important aspect of this paper however is the fact that we are able to calculate distinctive class of solutions which cannot be found in the current literature. In other words, using this new algebraic method the solution manifold is augmented to new class of solution functions. Simultaneously we would like to stress the necessity of such sophisticated methods since a general theory of nPDE does not exist. Otherwise, for practical use the algebraic construction of new class of solutions is of fundamental interest
Design techniques for large scale linear measurement systems
International Nuclear Information System (INIS)
Candy, J.V.
1979-03-01
Techniques to design measurement schemes for systems modeled by large scale linear time invariant systems, i.e., physical systems modeled by a large number (> 5) of ordinary differential equations, are described. The techniques are based on transforming the physical system model to a coordinate system facilitating the design and then transforming back to the original coordinates. An example of a three-stage, four-species, extraction column used in the reprocessing of spent nuclear fuel elements is presented. The basic ideas are briefly discussed in the case of noisy measurements. An example using a plutonium nitrate storage vessel (reprocessing) with measurement uncertainty is also presented
Li, Yuzhong
Using GA solve the winner determination problem (WDP) with large bids and items, run under different distribution, because the search space is large, constraint complex and it may easy to produce infeasible solution, would affect the efficiency and quality of algorithm. This paper present improved MKGA, including three operator: preprocessing, insert bid and exchange recombination, and use Monkey-king elite preservation strategy. Experimental results show that improved MKGA is better than SGA in population size and computation. The problem that traditional branch and bound algorithm hard to solve, improved MKGA can solve and achieve better effect.
Performance prediction of gas turbines by solving a system of non-linear equations
Energy Technology Data Exchange (ETDEWEB)
Kaikko, J
1998-09-01
This study presents a novel method for implementing the performance prediction of gas turbines from the component models. It is based on solving the non-linear set of equations that corresponds to the process equations, and the mass and energy balances for the engine. General models have been presented for determining the steady state operation of single components. Single and multiple shad arrangements have been examined with consideration also being given to heat regeneration and intercooling. Emphasis has been placed upon axial gas turbines of an industrial scale. Applying the models requires no information of the structural dimensions of the gas turbines. On comparison with the commonly applied component matching procedures, this method incorporates several advantages. The application of the models for providing results is facilitated as less attention needs to be paid to calculation sequences and routines. Solving the set of equations is based on zeroing co-ordinate functions that are directly derived from the modelling equations. Therefore, controlling the accuracy of the results is easy. This method gives more freedom for the selection of the modelling parameters since, unlike for the matching procedures, exchanging these criteria does not itself affect the algorithms. Implicit relationships between the variables are of no significance, thus increasing the freedom for the modelling equations as well. The mathematical models developed in this thesis will provide facilities to optimise the operation of any major gas turbine configuration with respect to the desired process parameters. The computational methods used in this study may also be adapted to any other modelling problems arising in industry. (orig.) 36 refs.
Linear differential equations to solve nonlinear mechanical problems: A novel approach
Nair, C. Radhakrishnan
2004-01-01
Often a non-linear mechanical problem is formulated as a non-linear differential equation. A new method is introduced to find out new solutions of non-linear differential equations if one of the solutions of a given non-linear differential equation is known. Using the known solution of the non-linear differential equation, linear differential equations are set up. The solutions of these linear differential equations are found using standard techniques. Then the solutions of the linear differe...
Research on geometric rectification of the Large FOV Linear Array Whiskbroom Image
Liu, Dia; Liu, Hui-tong; Dong, Hao; Liu, Xiao-bo
2015-08-01
To solve the geometric distortion problem of large FOV linear array whiskbroom image, a model of multi center central projection collinearity equation was founded considering its whiskbroom and linear CCD imaging feature, and the principle of distortion was analyzed. Based on the rectification method with POS, we introduced the angular position sensor data of the servo system, and restored the geometric imaging process exactly. An indirect rectification scheme aiming at linear array imaging with best scanline searching method was adopted, matrixes for calculating the exterior orientation elements was redesigned. We improved two iterative algorithms for this device, and did comparison and analysis. The rectification for the images of airborne imaging experiment showed ideal effect.
Large linear magnetoresistivity in strongly inhomogeneous planar and layered systems
International Nuclear Information System (INIS)
Bulgadaev, S.A.; Kusmartsev, F.V.
2005-01-01
Explicit expressions for magnetoresistance R of planar and layered strongly inhomogeneous two-phase systems are obtained, using exact dual transformation, connecting effective conductivities of in-plane isotropic two-phase systems with and without magnetic field. These expressions allow to describe the magnetoresistance of various inhomogeneous media at arbitrary concentrations x and magnetic fields H. All expressions show large linear magnetoresistance effect with different dependencies on the phase concentrations. The corresponding plots of the x- and H-dependencies of R(x,H) are represented for various values, respectively, of magnetic field and concentrations at some values of inhomogeneity parameter. The obtained results show a remarkable similarity with the existing experimental data on linear magnetoresistance in silver chalcogenides Ag 2+δ Se. A possible physical explanation of this similarity is proposed. It is shown that the random, stripe type, structures of inhomogeneities are the most suitable for a fabrication of magnetic sensors and a storage of information at room temperatures
Programmable Solution for Solving Non-linearity Characteristics of Smart Sensor Applications
Directory of Open Access Journals (Sweden)
S. Khan
2007-10-01
Full Text Available This paper presents a simple but programmable technique to solve the problem of non-linear characteristics of sensors used in more sensitive applications. The nonlinearity of the output response becomes a very sensitive issue in cases where a proportional increase in the physical quantity fails to bring about a proportional increase in the signal measured. The nonlinearity is addressed by using the interpolation method on the characteristics of a given sensor, approximating it to a set of tangent lines, the tangent points of which are recognized in the code of the processor by IF-THEN code. The method suggested here eliminates the use of external circuits for interfacing, and eases the programming burden on the processor at the cost of proportionally reduced memory requirements. The mathematically worked out results are compared with the simulation and experimental results for an IR sensor selected for the purpose and used for level measurement. This work will be of paramount importance and significance in applications where the controlled signal is required to follow the input signal precisely particularly in sensitive robotic applications.
Aihong Ren
2016-01-01
This paper is concerned with a class of fully fuzzy bilevel linear programming problems where all the coefficients and decision variables of both objective functions and the constraints are fuzzy numbers. A new approach based on deviation degree measures and a ranking function method is proposed to solve these problems. We first introduce concepts of the feasible region and the fuzzy optimal solution of a fully fuzzy bilevel linear programming problem. In order to obtain a fuzzy optimal solut...
Directory of Open Access Journals (Sweden)
Faridah Hani Mohamed Salleh
2017-01-01
Full Text Available Gene regulatory network (GRN reconstruction is the process of identifying regulatory gene interactions from experimental data through computational analysis. One of the main reasons for the reduced performance of previous GRN methods had been inaccurate prediction of cascade motifs. Cascade error is defined as the wrong prediction of cascade motifs, where an indirect interaction is misinterpreted as a direct interaction. Despite the active research on various GRN prediction methods, the discussion on specific methods to solve problems related to cascade errors is still lacking. In fact, the experiments conducted by the past studies were not specifically geared towards proving the ability of GRN prediction methods in avoiding the occurrences of cascade errors. Hence, this research aims to propose Multiple Linear Regression (MLR to infer GRN from gene expression data and to avoid wrongly inferring of an indirect interaction (A → B → C as a direct interaction (A → C. Since the number of observations of the real experiment datasets was far less than the number of predictors, some predictors were eliminated by extracting the random subnetworks from global interaction networks via an established extraction method. In addition, the experiment was extended to assess the effectiveness of MLR in dealing with cascade error by using a novel experimental procedure that had been proposed in this work. The experiment revealed that the number of cascade errors had been very minimal. Apart from that, the Belsley collinearity test proved that multicollinearity did affect the datasets used in this experiment greatly. All the tested subnetworks obtained satisfactory results, with AUROC values above 0.5.
Salleh, Faridah Hani Mohamed; Zainudin, Suhaila; Arif, Shereena M
2017-01-01
Gene regulatory network (GRN) reconstruction is the process of identifying regulatory gene interactions from experimental data through computational analysis. One of the main reasons for the reduced performance of previous GRN methods had been inaccurate prediction of cascade motifs. Cascade error is defined as the wrong prediction of cascade motifs, where an indirect interaction is misinterpreted as a direct interaction. Despite the active research on various GRN prediction methods, the discussion on specific methods to solve problems related to cascade errors is still lacking. In fact, the experiments conducted by the past studies were not specifically geared towards proving the ability of GRN prediction methods in avoiding the occurrences of cascade errors. Hence, this research aims to propose Multiple Linear Regression (MLR) to infer GRN from gene expression data and to avoid wrongly inferring of an indirect interaction (A → B → C) as a direct interaction (A → C). Since the number of observations of the real experiment datasets was far less than the number of predictors, some predictors were eliminated by extracting the random subnetworks from global interaction networks via an established extraction method. In addition, the experiment was extended to assess the effectiveness of MLR in dealing with cascade error by using a novel experimental procedure that had been proposed in this work. The experiment revealed that the number of cascade errors had been very minimal. Apart from that, the Belsley collinearity test proved that multicollinearity did affect the datasets used in this experiment greatly. All the tested subnetworks obtained satisfactory results, with AUROC values above 0.5.
Fonger, Nicole L.; Davis, Jon D.; Rohwer, Mary Lou
2018-01-01
This research addresses the issue of how to support students' representational fluency--the ability to create, move within, translate across, and derive meaning from external representations of mathematical ideas. The context of solving linear equations in a combined computer algebra system (CAS) and paper-and-pencil classroom environment is…
Solving and Interpreting Large-scale Harvest Scheduling Problems by Duality and Decomposition
Berck, Peter; Bible, Thomas
1982-01-01
This paper presents a solution to the forest planning problem that takes advantage of both the duality of linear programming formulations currently being used for harvest scheduling and the characteristics of decomposition inherent in the forest land class-relationship. The subproblems of decomposition, defined as the dual, can be solved in a simple, recursive fashion. In effect, such a technique reduces the computational burden in terms of time and computer storage as compared to the traditi...
Parallel Quasi Newton Algorithms for Large Scale Non Linear Unconstrained Optimization
International Nuclear Information System (INIS)
Rahman, M. A.; Basarudin, T.
1997-01-01
This paper discusses about Quasi Newton (QN) method to solve non-linear unconstrained minimization problems. One of many important of QN method is choice of matrix Hk. to be positive definite and satisfies to QN method. Our interest here is the parallel QN methods which will suite for the solution of large-scale optimization problems. The QN methods became less attractive in large-scale problems because of the storage and computational requirements. How ever, it is often the case that the Hessian is space matrix. In this paper we include the mechanism of how to reduce the Hessian update and hold the Hessian properties.One major reason of our research is that the QN method may be good in solving certain type of minimization problems, but it is efficiency degenerate when is it applied to solve other category of problems. For this reason, we use an algorithm containing several direction strategies which are processed in parallel. We shall attempt to parallelized algorithm by exploring different search directions which are generated by various QN update during the minimization process. The different line search strategies will be employed simultaneously in the process of locating the minimum along each direction.The code of algorithm will be written in Occam language 2 which is run on the transputer machine
A note on solving large-scale zero-one programming problems
Adema, Jos J.
1988-01-01
A heuristic for solving large-scale zero-one programming problems is provided. The heuristic is based on the modifications made by H. Crowder et al. (1983) to the standard branch-and-bound strategy. First, the initialization is modified. The modification is only useful if the objective function
Solving a large-scale precedence constrained scheduling problem with elastic jobs using tabu search
DEFF Research Database (Denmark)
Pedersen, C.R.; Rasmussen, R.V.; Andersen, Kim Allan
2007-01-01
exploitation of the elastic jobs and solve the problem using a tabu search procedure. Finding an initial feasible solution is in general -complete, but the tabu search procedure includes a specialized heuristic for solving this problem. The solution method has proven to be very efficient and leads......This paper presents a solution method for minimizing makespan of a practical large-scale scheduling problem with elastic jobs. The jobs are processed on three servers and restricted by precedence constraints, time windows and capacity limitations. We derive a new method for approximating the server...... to a significant decrease in makespan compared to the strategy currently implemented....
Solving a large-scale precedence constrained scheduling problem with elastic jobs using tabu search
DEFF Research Database (Denmark)
Pedersen, C.R.; Rasmussen, R.V.; Andersen, Kim Allan
2007-01-01
This paper presents a solution method for minimizing makespan of a practical large-scale scheduling problem with elastic jobs. The jobs are processed on three servers and restricted by precedence constraints, time windows and capacity limitations. We derive a new method for approximating the server...... exploitation of the elastic jobs and solve the problem using a tabu search procedure. Finding an initial feasible solution is in general -complete, but the tabu search procedure includes a specialized heuristic for solving this problem. The solution method has proven to be very efficient and leads...
DEFF Research Database (Denmark)
Amini Afshar, Mostafa; Bingham, Harry B.
2017-01-01
. Frequency-domain results are then obtained from a Fourier transform of the force and motion signals. In order to make a robust Fourier transform, and capture the response around the critical frequency, the tail of the force signal is asymptotically extrapolated assuming a linear decay rate. Fourth......The linearized potential flow approximation for the forward speed radiation problem is solved in the time domain using a high-order finite difference method. The finite-difference discretization is developed on overlapping, curvilinear body-fitted grids. To ensure numerical stability...
Solving large sets of coupled equations iteratively by vector processing on the CYBER 205 computer
International Nuclear Information System (INIS)
Tolsma, L.D.
1985-01-01
The set of coupled linear second-order differential equations which has to be solved for the quantum-mechanical description of inelastic scattering of atomic and nuclear particles can be rewritten as an equivalent set of coupled integral equations. When some type of functions is used as piecewise analytic reference solutions, the integrals that arise in this set can be evaluated analytically. The set of integral equations can be solved iteratively. For the results mentioned an inward-outward iteration scheme has been applied. A concept of vectorization of coupled-channel Fortran programs, based on this integral method, is presented for the use on the Cyber 205 computer. It turns out that, for two heavy ion nuclear scattering test cases, this vector algorithm gives an overall speed-up of about a factor of 2 to 3 compared to a highly optimized scalar algorithm for a one vector pipeline computer
Yanti, Y. R.; Amin, S. M.; Sulaiman, R.
2018-01-01
This study described representation of students who have musical, logical-mathematic and naturalist intelligence in solving a problem. Subjects were selected on the basis of multiple intelligence tests (TPM) consists of 108 statements, with 102 statements adopted from Chislet and Chapman and 6 statements equal to eksistensial intelligences. Data were analyzed based on problem-solving tests (TPM) and interviewing. See the validity of the data then problem-solving tests (TPM) and interviewing is given twice with an analyzed using the representation indikator and the problem solving step. The results showed that: the stage of presenting information known, stage of devising a plan, and stage of carrying out the plan those three subjects were using same form of representation. While he stage of presenting information asked and stage of looking back, subject of logical-mathematic was using different forms of representation with subjects of musical and naturalist intelligence. From this research is expected to provide input to the teacher in determining the learning strategy that will be used by considering the representation of students with the basis of multiple intelligences.
Directory of Open Access Journals (Sweden)
Samir Dey
2015-07-01
Full Text Available This paper proposes a new multi-objective intuitionistic fuzzy goal programming approach to solve a multi-objective nonlinear programming problem in context of a structural design. Here we describe some basic properties of intuitionistic fuzzy optimization. We have considered a multi-objective structural optimization problem with several mutually conflicting objectives. The design objective is to minimize weight of the structure and minimize the vertical deflection at loading point of a statistically loaded three-bar planar truss subjected to stress constraints on each of the truss members. This approach is used to solve the above structural optimization model based on arithmetic mean and compare with the solution by intuitionistic fuzzy goal programming approach. A numerical solution is given to illustrate our approach.
Directory of Open Access Journals (Sweden)
Shahid Hasnain
2017-07-01
Full Text Available This research paper represents a numerical approximation to non-linear three dimension reaction diffusion equation with non-linear source term from population genetics. Since various initial and boundary value problems exist in three dimension reaction diffusion phenomena, which are studied numerically by different numerical methods, here we use finite difference schemes (Alternating Direction Implicit and Fourth Order Douglas Implicit to approximate the solution. Accuracy is studied in term of L2, L∞ and relative error norms by random selected grids along time levels for comparison with analytical results. The test example demonstrates the accuracy, efficiency and versatility of the proposed schemes. Numerical results showed that Fourth Order Douglas Implicit scheme is very efficient and reliable for solving 3-D non-linear reaction diffusion equation.
Hasnain, Shahid; Saqib, Muhammad; Mashat, Daoud Suleiman
2017-07-01
This research paper represents a numerical approximation to non-linear three dimension reaction diffusion equation with non-linear source term from population genetics. Since various initial and boundary value problems exist in three dimension reaction diffusion phenomena, which are studied numerically by different numerical methods, here we use finite difference schemes (Alternating Direction Implicit and Fourth Order Douglas Implicit) to approximate the solution. Accuracy is studied in term of L2, L∞ and relative error norms by random selected grids along time levels for comparison with analytical results. The test example demonstrates the accuracy, efficiency and versatility of the proposed schemes. Numerical results showed that Fourth Order Douglas Implicit scheme is very efficient and reliable for solving 3-D non-linear reaction diffusion equation.
Large-scale dynamo action due to α fluctuations in a linear shear flow
Sridhar, S.; Singh, Nishant K.
2014-12-01
We present a model of large-scale dynamo action in a shear flow that has stochastic, zero-mean fluctuations of the α parameter. This is based on a minimal extension of the Kraichnan-Moffatt model, to include a background linear shear and Galilean-invariant α-statistics. Using the first-order smoothing approximation we derive a linear integro-differential equation for the large-scale magnetic field, which is non-perturbative in the shearing rate S , and the α-correlation time τα . The white-noise case, τα = 0 , is solved exactly, and it is concluded that the necessary condition for dynamo action is identical to the Kraichnan-Moffatt model without shear; this is because white-noise does not allow for memory effects, whereas shear needs time to act. To explore memory effects we reduce the integro-differential equation to a partial differential equation, valid for slowly varying fields when τα is small but non-zero. Seeking exponential modal solutions, we solve the modal dispersion relation and obtain an explicit expression for the growth rate as a function of the six independent parameters of the problem. A non-zero τα gives rise to new physical scales, and dynamo action is completely different from the white-noise case; e.g. even weak α fluctuations can give rise to a dynamo. We argue that, at any wavenumber, both Moffatt drift and Shear always contribute to increasing the growth rate. Two examples are presented: (a) a Moffatt drift dynamo in the absence of shear and (b) a Shear dynamo in the absence of Moffatt drift.
A neural network method for solving a system of linear variational inequalities
International Nuclear Information System (INIS)
Lan Hengyou; Cui Yishun
2009-01-01
In this paper, we transmute the solution for a new system of linear variational inequalities to an equilibrium point of neural networks, and by using analytic technique, some sufficient conditions are presented. Further, the estimation of the exponential convergence rates of the neural networks is investigated. The new and useful results obtained in this paper generalize and improve the corresponding results of recent works.
Lancellotti, V.; Tijhuis, A.G.
2012-01-01
The calculation of electromagnetic (EM) fields and waves inside finite-sized structures comprised of different media can benefit from a diakoptics method such as linear embedding via Green's operators (LEGO). Unlike scattering problems, the excitation of EM waves within the bulk dielectric requires
An Interactive Method to Solve Infeasibility in Linear Programming Test Assembling Models
Huitzing, Hiddo A.
2004-01-01
In optimal assembly of tests from item banks, linear programming (LP) models have proved to be very useful. Assembly by hand has become nearly impossible, but these LP techniques are able to find the best solutions, given the demands and needs of the test to be assembled and the specifics of the item bank from which it is assembled. However,…
Solving Large-Scale Computational Problems Using Insights from Statistical Physics
Energy Technology Data Exchange (ETDEWEB)
Selman, Bart [Cornell University
2012-02-29
Many challenging problems in computer science and related fields can be formulated as constraint satisfaction problems. Such problems consist of a set of discrete variables and a set of constraints between those variables, and represent a general class of so-called NP-complete problems. The goal is to find a value assignment to the variables that satisfies all constraints, generally requiring a search through and exponentially large space of variable-value assignments. Models for disordered systems, as studied in statistical physics, can provide important new insights into the nature of constraint satisfaction problems. Recently, work in this area has resulted in the discovery of a new method for solving such problems, called the survey propagation (SP) method. With SP, we can solve problems with millions of variables and constraints, an improvement of two orders of magnitude over previous methods.
A large superconducting accelerator project. International linear collider (ILC). Introduction
International Nuclear Information System (INIS)
Yamamoto, Akira
2013-01-01
The international linear collider (ILC) is proposed as the next-energy-frontier particle accelerator anticipated to be realized through global cooperation. The ILC accelerator is composed of a pair of electron and positron linear accelerators to realize head-on collision with a center-of-mass energy of 500 (250+250) GeV. It is based on superconducting radio-frequency (SCRF) technology, and the R and D and technical design have progressed in the technical design phase since 2007, and the technical design report (TDR) reached completion in 2012. This report reviews the ILC general design and technology. (author)
A comparison of iterative methods to solve complex valued linear algebraic systems
Czech Academy of Sciences Publication Activity Database
Axelsson, Owe; Neytcheva, M.; Ahmad, B.
2013-01-01
Roč. 66, č. 4 (2013), s. 811-841 ISSN 1017-1398 R&D Projects: GA MŠk ED1.1.00/02.0070 Institutional support: RVO:68145535 Keywords : linear systems * complex symmetric * real valued form * preconditioning Subject RIV: BA - General Mathematics Impact factor: 1.005, year: 2013 http://www.it.uu.se/research/publications/reports/2013-005/2013-005-nc.pdf
An Improved Search Approach for Solving Non-Convex Mixed-Integer Non Linear Programming Problems
Sitopu, Joni Wilson; Mawengkang, Herman; Syafitri Lubis, Riri
2018-01-01
The nonlinear mathematical programming problem addressed in this paper has a structure characterized by a subset of variables restricted to assume discrete values, which are linear and separable from the continuous variables. The strategy of releasing nonbasic variables from their bounds, combined with the “active constraint” method, has been developed. This strategy is used to force the appropriate non-integer basic variables to move to their neighbourhood integer points. Successful implementation of these algorithms was achieved on various test problems.
A frequency-domain method for solving linear time delay systems with constant coefficients
Jin, Mengshi; Chen, Wei; Song, Hanwen; Xu, Jian
2018-03-01
In an active control system, time delay will occur due to processes such as signal acquisition and transmission, calculation, and actuation. Time delay systems are usually described by delay differential equations (DDEs). Since it is hard to obtain an analytical solution to a DDE, numerical solution is of necessity. This paper presents a frequency-domain method that uses a truncated transfer function to solve a class of DDEs. The theoretical transfer function is the sum of infinite items expressed in terms of poles and residues. The basic idea is to select the dominant poles and residues to truncate the transfer function, thus ensuring the validity of the solution while improving the efficiency of calculation. Meanwhile, the guideline of selecting these poles and residues is provided. Numerical simulations of both stable and unstable delayed systems are given to verify the proposed method, and the results are presented and analysed in detail.
Effective quadrature formula in solving linear integro-differential equations of order two
Eshkuvatov, Z. K.; Kammuji, M.; Long, N. M. A. Nik; Yunus, Arif A. M.
2017-08-01
In this note, we solve general form of Fredholm-Volterra integro-differential equations (IDEs) of order 2 with boundary condition approximately and show that proposed method is effective and reliable. Initially, IDEs is reduced into integral equation of the third kind by using standard integration techniques and identity between multiple and single integrals then truncated Legendre series are used to estimate the unknown function. For the kernel integrals, we have applied Gauss-Legendre quadrature formula and collocation points are chosen as the roots of the Legendre polynomials. Finally, reduce the integral equations of the third kind into the system of algebraic equations and Gaussian elimination method is applied to get approximate solutions. Numerical examples and comparisons with other methods reveal that the proposed method is very effective and dominated others in many cases. General theory of existence of the solution is also discussed.
He, Qiang; Hu, Xiangtao; Ren, Hong; Zhang, Hongqi
2015-11-01
A novel artificial fish swarm algorithm (NAFSA) is proposed for solving large-scale reliability-redundancy allocation problem (RAP). In NAFSA, the social behaviors of fish swarm are classified in three ways: foraging behavior, reproductive behavior, and random behavior. The foraging behavior designs two position-updating strategies. And, the selection and crossover operators are applied to define the reproductive ability of an artificial fish. For the random behavior, which is essentially a mutation strategy, the basic cloud generator is used as the mutation operator. Finally, numerical results of four benchmark problems and a large-scale RAP are reported and compared. NAFSA shows good performance in terms of computational accuracy and computational efficiency for large scale RAP. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.
Solving the linear radiation problem using a volume method on an overset grid
DEFF Research Database (Denmark)
Read, Robert; Bingham, Harry B.
2012-01-01
of numerical results with established analytical solutions. The linear radiation problem is considered in this paper. A two-dimensional computational tool has been developed to calculate the force applied to a floating body of arbitrary form in response to a prescribed displacement. Fourier transforms......This paper describes recent progress towards the development of a computational tool, based on potential ow theory, that can accurately and effciently simulate wave-induced loadings on marine structures. Engsig-Karup et al. (2009) have successfully developed an arbitrary-order, finite...
On Numerical Stability in Large Scale Linear Algebraic Computations
Czech Academy of Sciences Publication Activity Database
Strakoš, Zdeněk; Liesen, J.
2005-01-01
Roč. 85, č. 5 (2005), s. 307-325 ISSN 0044-2267 R&D Projects: GA AV ČR 1ET400300415 Institutional research plan: CEZ:AV0Z10300504 Keywords : linear algebraic systems * eigenvalue problems * convergence * numerical stability * backward error * accuracy * Lanczos method * conjugate gradient method * GMRES method Subject RIV: BA - General Mathematics Impact factor: 0.351, year: 2005
International Nuclear Information System (INIS)
Secher, Bernard; Belliard, Michel; Calvin, Christophe
2009-01-01
This paper describes a tool called 'Numerical Platon' developed by the French Atomic Energy Commission (CEA). It provides a freely available (GNU LGPL license) interface for coupling scientific computing applications to various freeware linear solver libraries (essentially PETSc, SuperLU and HyPre), together with some proprietary CEA solvers, for high-performance computers that may be used in industrial software written in various programming languages. This tool was developed as part of considerable efforts by the CEA Nuclear Energy Division in the past years to promote massively parallel software and on-shelf parallel tools to help develop new generation simulation codes. After the presentation of the package architecture and the available algorithms, we show examples of how Numerical Platon is used in sequential and parallel CEA codes. Comparing with in-house solvers, the gain in terms of increases in computation capacities or in terms of parallel performances is notable, without considerable extra development cost
hi-class: Horndeski in the Cosmic Linear Anisotropy Solving System
Energy Technology Data Exchange (ETDEWEB)
Zumalacárregui, Miguel [Nordita, KHT Royal Institute of Technology and Stockholm University, Roslagstullsbacken 23, SE-106 91 Stockholm (Sweden); Bellini, Emilio [Institut de Ciènces del Cosmos, Universitat de Barcelona, IEEC-UB, Martì i Franquè 1, E-08028 Barcelona (Spain); Sawicki, Ignacy [Central European Institute for Cosmology and Fundamental Physics, Fyzikální ustáv Akademie v\\v ed \\v CR, Na Slovance 2, 182 21 Praha 8 (Czech Republic); Lesgourgues, Julien [Institut für Theoretische Teilchenphysik und Kosmologie, RWTH Aachen University, D-52056 Aachen (Germany); Ferreira, Pedro G., E-mail: miguelzuma@berkeley.edu, E-mail: emilio.bellini@physics.ox.ac.uk, E-mail: ignacy.sawicki@fzu.cz, E-mail: lesgourg@physik.rwth-aachen.de, E-mail: pedro.ferreira@physics.ox.ac.uk [Astrophysics, University of Oxford, Denys Wilkinson Building, Keble Road, Oxford OX1 3RH (United Kingdom)
2017-08-01
We present the public version of hi-class (www.hiclass-code.net), an extension of the Boltzmann code CLASS to a broad ensemble of modifications to general relativity. In particular, hi-class can calculate predictions for models based on Horndeski's theory, which is the most general scalar-tensor theory described by second-order equations of motion and encompasses any perfect-fluid dark energy, quintessence, Brans-Dicke, f ( R ) and covariant Galileon models. hi-class has been thoroughly tested and can be readily used to understand the impact of alternative theories of gravity on linear structure formation as well as for cosmological parameter extraction.
Energy Technology Data Exchange (ETDEWEB)
Secher, Bernard [French Atomic Energy Commission (CEA), Nuclear Energy Division (DEN) (France); CEA Saclay DM2S/SFME/LGLS, Bat. 454, F-91191 Gif-sur-Yvette Cedex (France)], E-mail: bsecher@cea.fr; Belliard, Michel [French Atomic Energy Commission (CEA), Nuclear Energy Division (DEN) (France); CEA Cadarache DER/SSTH/LMDL, Bat. 238, F-13108 Saint-Paul-lez-Durance Cedex (France); Calvin, Christophe [French Atomic Energy Commission (CEA), Nuclear Energy Division (DEN) (France); CEA Saclay DM2S/SERMA/LLPR, Bat. 470, F-91191 Gif-sur-Yvette Cedex (France)
2009-01-15
This paper describes a tool called 'Numerical Platon' developed by the French Atomic Energy Commission (CEA). It provides a freely available (GNU LGPL license) interface for coupling scientific computing applications to various freeware linear solver libraries (essentially PETSc, SuperLU and HyPre), together with some proprietary CEA solvers, for high-performance computers that may be used in industrial software written in various programming languages. This tool was developed as part of considerable efforts by the CEA Nuclear Energy Division in the past years to promote massively parallel software and on-shelf parallel tools to help develop new generation simulation codes. After the presentation of the package architecture and the available algorithms, we show examples of how Numerical Platon is used in sequential and parallel CEA codes. Comparing with in-house solvers, the gain in terms of increases in computation capacities or in terms of parallel performances is notable, without considerable extra development cost.
Constraints to solve parallelogram grid problems in 2D non separable linear canonical transform
Zhao, Liang; Healy, John J.; Muniraj, Inbarasan; Cui, Xiao-Guang; Malallah, Ra'ed; Ryle, James P.; Sheridan, John T.
2017-05-01
The 2D non-separable linear canonical transform (2D-NS-LCT) can model a range of various paraxial optical systems. Digital algorithms to evaluate the 2D-NS-LCTs are important in modeling the light field propagations and also of interest in many digital signal processing applications. In [Zhao 14] we have reported that a given 2D input image with rectangular shape/boundary, in general, results in a parallelogram output sampling grid (generally in an affine coordinates rather than in a Cartesian coordinates) thus limiting the further calculations, e.g. inverse transform. One possible solution is to use the interpolation techniques; however, it reduces the speed and accuracy of the numerical approximations. To alleviate this problem, in this paper, some constraints are derived under which the output samples are located in the Cartesian coordinates. Therefore, no interpolation operation is required and thus the calculation error can be significantly eliminated.
DEFF Research Database (Denmark)
Sorokin, Vladislav; Thomsen, Jon Juel
2015-01-01
Parametrically excited systems appear in many fields of science and technology, intrinsically or imposed purposefully; e.g. spatially periodic structures represent an important class of such systems [4]. When the parametric excitation can be considered weak, classical asymptotic methods like...... the method of averaging [2] or multiple scales [6] can be applied. However, with many practically important applications this simplification is inadequate, e.g. with spatially periodic structures it restricts the possibility to affect their effective dynamic properties by a structural parameter modulation...... of considerable magnitude. Approximate methods based on Floquet theory [4] for analyzing problems involving parametric excitation, e.g. the classical Hill’s method of infinite determinants [3,4], can be employed also in cases of strong excitation; however, with Floquet theory being applicable only for linear...
Solving Large-Scale TSP Using a Fast Wedging Insertion Partitioning Approach
Directory of Open Access Journals (Sweden)
Zuoyong Xiang
2015-01-01
Full Text Available A new partitioning method, called Wedging Insertion, is proposed for solving large-scale symmetric Traveling Salesman Problem (TSP. The idea of our proposed algorithm is to cut a TSP tour into four segments by nodes’ coordinate (not by rectangle, such as Strip, FRP, and Karp. Each node is located in one of their segments, which excludes four particular nodes, and each segment does not twist with other segments. After the partitioning process, this algorithm utilizes traditional construction method, that is, the insertion method, for each segment to improve the quality of tour, and then connects the starting node and the ending node of each segment to obtain the complete tour. In order to test the performance of our proposed algorithm, we conduct the experiments on various TSPLIB instances. The experimental results show that our proposed algorithm in this paper is more efficient for solving large-scale TSPs. Specifically, our approach is able to obviously reduce the time complexity for running the algorithm; meanwhile, it will lose only about 10% of the algorithm’s performance.
Robust linear discriminant models to solve financial crisis in banking sectors
Lim, Yai-Fung; Yahaya, Sharipah Soaad Syed; Idris, Faoziah; Ali, Hazlina; Omar, Zurni
2014-12-01
Linear discriminant analysis (LDA) is a widely-used technique in patterns classification via an equation which will minimize the probability of misclassifying cases into their respective categories. However, the performance of classical estimators in LDA highly depends on the assumptions of normality and homoscedasticity. Several robust estimators in LDA such as Minimum Covariance Determinant (MCD), S-estimators and Minimum Volume Ellipsoid (MVE) are addressed by many authors to alleviate the problem of non-robustness of the classical estimates. In this paper, we investigate on the financial crisis of the Malaysian banking institutions using robust LDA and classical LDA methods. Our objective is to distinguish the "distress" and "non-distress" banks in Malaysia by using the LDA models. Hit ratio is used to validate the accuracy predictive of LDA models. The performance of LDA is evaluated by estimating the misclassification rate via apparent error rate. The results and comparisons show that the robust estimators provide a better performance than the classical estimators for LDA.
Yager’s ranking method for solving the trapezoidal fuzzy number linear programming
Karyati; Wutsqa, D. U.; Insani, N.
2018-03-01
In the previous research, the authors have studied the fuzzy simplex method for trapezoidal fuzzy number linear programming based on the Maleki’s ranking function. We have found some theories related to the term conditions for the optimum solution of fuzzy simplex method, the fuzzy Big-M method, the fuzzy two-phase method, and the sensitivity analysis. In this research, we study about the fuzzy simplex method based on the other ranking function. It is called Yager's ranking function. In this case, we investigate the optimum term conditions. Based on the result of research, it is found that Yager’s ranking function is not like Maleki’s ranking function. Using the Yager’s function, the simplex method cannot work as well as when using the Maleki’s function. By using the Yager’s function, the value of the subtraction of two equal fuzzy numbers is not equal to zero. This condition makes the optimum table of the fuzzy simplex table is undetected. As a result, the simplified fuzzy simplex table becomes stopped and does not reach the optimum solution.
Liu, Tao; Huang, Jie
2017-04-17
This paper presents a discrete-time recurrent neural network approach to solving systems of linear equations with two features. First, the system of linear equations may not have a unique solution. Second, the system matrix is not known precisely, but a sequence of matrices that converges to the unknown system matrix exponentially is known. The problem is motivated from solving the output regulation problem for linear systems. Thus, an application of our main result leads to an online solution to the output regulation problem for linear systems.
International Nuclear Information System (INIS)
Murfi, Hendri; Basaruddin, T.
2001-01-01
The interior point method for linear programming has gained extraordinary interest as an alternative to simplex method since Karmarkar presented a polynomial-time algorithm for linear programming based on interior point method. In implementation of the algorithm of this method, there are two important things that have impact heavily to performance of the algorithm; they are data structure and used method to solve linear equation system in the algorithm. This paper describes about solving linear equation system in variants of the algorithm called dual-affine scaling algorithm. Next, we evaluate experimentally results of some used methods, either direct method or iterative method. The experimental evaluation used Matlab
Gross, Lutz; Altinay, Cihan; Fenwick, Joel; Smith, Troy
2014-05-01
inversion and appropriate solution schemes in escript. We will also give a brief introduction into escript's open framework for defining and solving geophysical inversion problems. Finally we will show some benchmark results to demonstrate the computational scalability of the inversion method across a large number of cores and compute nodes in a parallel computing environment. References: - L. Gross et al. (2013): Escript Solving Partial Differential Equations in Python Version 3.4, The University of Queensland, https://launchpad.net/escript-finley - L. Gross and C. Kemp (2013) Large Scale Joint Inversion of Geophysical Data using the Finite Element Method in escript. ASEG Extended Abstracts 2013, http://dx.doi.org/10.1071/ASEG2013ab306 - T. Poulet, L. Gross, D. Georgiev, J. Cleverley (2012): escript-RT: Reactive transport simulation in Python using escript, Computers & Geosciences, Volume 45, 168-176. http://dx.doi.org/10.1016/j.cageo.2011.11.005.
hi_class: Horndeski in the cosmic linear anisotropy solving system
Czech Academy of Sciences Publication Activity Database
Zumalacarregui, M.; Bellini, E.; Sawicki, Ignacy; Lesgourgues, J.; Ferreira, P.G.
2017-01-01
Roč. 2017, č. 8 (2017), s. 1-29, č. článku 019. ISSN 1475-7516 R&D Projects: GA MŠk EF15_003/0000437 Grant - others:OP VVV - CoGraDS(XE) CZ.02.1.01/0.0/0.0/15_003/0000437 Institutional support: RVO:68378271 Keywords : modified gravity * gravitational waves * cosmology * large scale structure Subject RIV: BN - Astronomy, Celestial Mechanics, Astrophysics OBOR OECD: Astronomy (including astrophysics,space science) Impact factor: 4.734, year: 2016
Penders, Bart; Vos, Rein; Horstman, Klasien
2009-11-01
Solving complex problems in large-scale research programmes requires cooperation and division of labour. Simultaneously, large-scale problem solving also gives rise to unintended side effects. Based upon 5 years of researching two large-scale nutrigenomic research programmes, we argue that problems are fragmented in order to be solved. These sub-problems are given priority for practical reasons and in the process of solving them, various changes are introduced in each sub-problem. Combined with additional diversity as a result of interdisciplinarity, this makes reassembling the original and overall goal of the research programme less likely. In the case of nutrigenomics and health, this produces a diversification of health. As a result, the public health goal of contemporary nutrition science is not reached in the large-scale research programmes we studied. Large-scale research programmes are very successful in producing scientific publications and new knowledge; however, in reaching their political goals they often are less successful.
Directory of Open Access Journals (Sweden)
Aihong Ren
2016-01-01
Full Text Available This paper is concerned with a class of fully fuzzy bilevel linear programming problems where all the coefficients and decision variables of both objective functions and the constraints are fuzzy numbers. A new approach based on deviation degree measures and a ranking function method is proposed to solve these problems. We first introduce concepts of the feasible region and the fuzzy optimal solution of a fully fuzzy bilevel linear programming problem. In order to obtain a fuzzy optimal solution of the problem, we apply deviation degree measures to deal with the fuzzy constraints and use a ranking function method of fuzzy numbers to rank the upper and lower level fuzzy objective functions. Then the fully fuzzy bilevel linear programming problem can be transformed into a deterministic bilevel programming problem. Considering the overall balance between improving objective function values and decreasing allowed deviation degrees, the computational procedure for finding a fuzzy optimal solution is proposed. Finally, a numerical example is provided to illustrate the proposed approach. The results indicate that the proposed approach gives a better optimal solution in comparison with the existing method.
High Precision Survey and Alignment of Large Linear Accelerators
Prenting, J
2004-01-01
For the future linear accelerator TESLA the demanded accuracy for the alignment of the components is 0.5 mm horizontal and 0.2 mm vertical, both on each 600 m section. Other accelerators require similar accuracies. These demands can not be fulfilled with open-air geodetic methods, mainly because of refraction. Therefore the RTRS (Rapid Tunnel Reference Surveyor), a measurement train performing overlapping multipoint alignment on a reference network is being developed. Two refraction-free realizations of this concept are being developed at the moment: the first one (GeLiS) measures the horizontal co-ordinates using stretched wires, combined with photogrammetric split-image sensors in a distance measurement configuration. In areas of the tunnel where the accelerator is following the earth curvature GeLiS measures the height using a new hydrostatic leveling system. The second concept (LiCAS) is based on laser straightness monitors (LSM) combined with frequency scanning interferometry (FSI) in an evacuated system...
Energy Technology Data Exchange (ETDEWEB)
Oliva, Jose de Jesus Rivero, E-mail: rivero@con.ufrj.br [Universidade Federal do Rio de Janeiro (UFRJ), RJ (Brazil). Departamento de Engenharia Nuclear; Sousa, Fernando Nuno Carneiro de, E-mail: fernandonunosousa@gmail.com [Aceletron Irradiacao lndustrial, Rio de Janeiro, RJ (Brazil)
2013-07-01
The electron linear accelerators can be used in industrial applications that require powerful sources of ionizing radiation. They have the important characteristic of not representing a radiation hazard when the accelerators remain electrically disconnected. With the plant in operation, a high reliability defense in depth reduces the risk of radiological accidents to extremely small levels. It is practically impossible that a person could enter into the radiation bunker with the accelerators connected. Aceletron Irradiacao Industrial, located in Rio de Janeiro, offers services of irradiation by means of two powerful electron linear accelerators, with 15 kW power and 10 MeV electron energy. Despite the high level of existing radiation safety, a simplified risk study is underway to identify possible sequences of radiological accidents. The study is based on the combined application of the event and fault trees techniques. Preliminary results confirm that there is a very small risk of entering into the irradiation bunker with the accelerators in operation, but the risk of an operator entering into the bunker during a process interruption and remaining there without notice after the accelerators were restarted may be considerably larger. Based on these results the Company is considering alternatives to reduce the likelihood of human error of this type that could lead to a radiological accident. The paper describes the defense in depth of the irradiation process in Aceletron Irradiacao Industrial, as well as the models and preliminary results of the ongoing risk analysis, including the additional safety measures which are being evaluated. (author)
Directory of Open Access Journals (Sweden)
Salvador Lucas
2015-12-01
Full Text Available Recent developments in termination analysis for declarative programs emphasize the use of appropriate models for the logical theory representing the program at stake as a generic approach to prove termination of declarative programs. In this setting, Order-Sorted First-Order Logic provides a powerful framework to represent declarative programs. It also provides a target logic to obtain models for other logics via transformations. We investigate the automatic generation of numerical models for order-sorted first-order logics and its use in program analysis, in particular in termination analysis of declarative programs. We use convex domains to give domains to the different sorts of an order-sorted signature; we interpret the ranked symbols of sorted signatures by means of appropriately adapted convex matrix interpretations. Such numerical interpretations permit the use of existing algorithms and tools from linear algebra and arithmetic constraint solving to synthesize the models.
Directory of Open Access Journals (Sweden)
H Kazemipoor
2012-04-01
Full Text Available A multi-skilled project scheduling problem (MSPSP has been generally presented to schedule a project with staff members as resources. Each activity in project network requires different skills and also staff members have different skills, too. This causes the MSPSP becomes a special type of a multi-mode resource-constrained project scheduling problem (MM-RCPSP with a huge number of modes. Given the importance of this issue, in this paper, a mixed integer linear programming for the MSPSP is presented. Due to the complexity of the problem, a meta-heuristic algorithm is proposed in order to find near optimal solutions. To validate performance of the algorithm, results are compared against exact solutions solved by the LINGO solver. The results are promising and show that optimal or near-optimal solutions are derived for small instances and good solutions for larger instances in reasonable time.
Energy Technology Data Exchange (ETDEWEB)
de la Torre Vega, E. [Instituto de Investigaciones Electricas, Cuernavaca (Mexico); Cesar Suarez Arriaga, M. [Universidad Michoacana SNH, Michoacan (Mexico)
1995-03-01
In geothermal simulation processes, MULKOM uses Integrated Finite Differences to solve the corresponding partial differential equations. This method requires to resolve efficiently big linear dispersed systems of non-symmetrical nature on each temporal iteration. The order of the system is usually greater than one thousand its solution could represent around 80% of CPU total calculation time. If the elapsed time solving this class of linear systems is reduced, the duration of numerical simulation decreases notably. When the matrix is big (N{ge}500) and with holes, it is inefficient to handle all the system`s elements, because it is perfectly figured out by its elements distinct of zero, quantity greatly minor than N{sup 2}. In this area, iteration methods introduce advantages with respect to gaussian elimination methods, because these last replenish matrices not having any special distribution of their non-zero elements and because they do not make use of the available solution estimations. The iterating methods of the Conjugated Gradient family, based on the subspaces of Krylov, possess the advantage of improving the convergence speed by means of preconditioning techniques. The creation of DIOMRES(k,m) method guarantees the continuous descent of the residual norm, without incurring in division by zero. This technique converges at most in N iterations if the system`s matrix is symmetrical, it does not employ too much memory to converge and updates immediately the approximation by using incomplete orthogonalization and adequate restarting. A preconditioned version of DIOMRES was applied to problems related to unsymmetrical systems with 1000 unknowns and less than five terms per equation. We found that this technique could reduce notably the time needful to find the solution without requiring memory increment. The coupling of this method to geothermal versions of MULKOM is in process.
Decentralised stabilising controllers for a class of large-scale linear ...
Indian Academy of Sciences (India)
subsystems resulting from a new aggregation-decomposition technique. The method has been illustrated through a numerical example of a large-scale linear system consisting of three subsystems each of the fourth order. Keywords. Decentralised stabilisation; large-scale linear systems; optimal feedback control; algebraic ...
Solving large scale unit dilemma in electricity system by applying commutative law
Legino, Supriadi; Arianto, Rakhmat
2018-03-01
The conventional system, pooling resources with large centralized power plant interconnected as a network. provides a lot of advantages compare to the isolated one include optimizing efficiency and reliability. However, such a large plant need a huge capital. In addition, more problems emerged to hinder the construction of big power plant as well as its associated transmission lines. By applying commutative law of math, ab = ba, for all a,b €-R, the problem associated with conventional system as depicted above, can be reduced. The idea of having small unit but many power plants, namely “Listrik Kerakyatan,” abbreviated as LK provides both social and environmental benefit that could be capitalized by using proper assumption. This study compares the cost and benefit of LK to those of conventional system, using simulation method to prove that LK offers alternative solution to answer many problems associated with the large system. Commutative Law of Algebra can be used as a simple mathematical model to analyze whether the LK system as an eco-friendly distributed generation can be applied to solve various problems associated with a large scale conventional system. The result of simulation shows that LK provides more value if its plants operate in less than 11 hours as peaker power plant or load follower power plant to improve load curve balance of the power system. The result of simulation indicates that the investment cost of LK plant should be optimized in order to minimize the plant investment cost. This study indicates that the benefit of economies of scale principle does not always apply to every condition, particularly if the portion of intangible cost and benefit is relatively high.
Liu, Yang
2013-07-01
The computational complexity and memory requirements of multilevel plane wave time domain (PWTD)-accelerated marching-on-in-time (MOT)-based surface integral equation (SIE) solvers scale as O(NtNs(log 2)Ns) and O(Ns 1.5); here N t and Ns denote numbers of temporal and spatial basis functions discretizing the current [Shanker et al., IEEE Trans. Antennas Propag., 51, 628-641, 2003]. In the past, serial versions of these solvers have been successfully applied to the analysis of scattering from perfect electrically conducting as well as homogeneous penetrable targets involving up to Ns ≈ 0.5 × 106 and Nt ≈ 10 3. To solve larger problems, parallel PWTD-enhanced MOT solvers are called for. Even though a simple parallelization strategy was demonstrated in the context of electromagnetic compatibility analysis [M. Lu et al., in Proc. IEEE Int. Symp. AP-S, 4, 4212-4215, 2004], by and large, progress in this area has been slow. The lack of progress can be attributed wholesale to difficulties associated with the construction of a scalable PWTD kernel. © 2013 IEEE.
Decomposition and parallelization strategies for solving large-scale MDO problems
Energy Technology Data Exchange (ETDEWEB)
Grauer, M.; Eschenauer, H.A. [Research Center for Multidisciplinary Analyses and Applied Structural Optimization, FOMAAS, Univ. of Siegen (Germany)
2007-07-01
During previous years, structural optimization has been recognized as a useful tool within the discriptiones of engineering and economics. However, the optimization of large-scale systems or structures is impeded by an immense solution effort. This was the reason to start a joint research and development (R and D) project between the Institute of Mechanics and Control Engineering and the Information and Decision Sciences Institute within the Research Center for Multidisciplinary Analyses and Applied Structural Optimization (FOMAAS) on cluster computing for parallel and distributed solution of multidisciplinary optimization (MDO) problems based on the OpTiX-Workbench. Here the focus of attention will be put on coarsegrained parallelization and its implementation on clusters of workstations. A further point of emphasis was laid on the development of a parallel decomposition strategy called PARDEC, for the solution of very complex optimization problems which cannot be solved efficiently by sequential integrated optimization. The use of the OptiX-Workbench together with the FEM ground water simulation system FEFLOW is shown for a special water management problem. (orig.)
Energy Technology Data Exchange (ETDEWEB)
Cobb, J.W.
1995-02-01
There is an increasing need for more accurate numerical methods for large-scale nonlinear magneto-fluid turbulence calculations. These methods should not only increase the current state of the art in terms of accuracy, but should also continue to optimize other desired properties such as simplicity, minimized computation, minimized memory requirements, and robust stability. This includes the ability to stably solve stiff problems with long time-steps. This work discusses a general methodology for deriving higher-order numerical methods. It also discusses how the selection of various choices can affect the desired properties. The explicit discussion focuses on third-order Runge-Kutta methods, including general solutions and five examples. The study investigates the linear numerical analysis of these methods, including their accuracy, general stability, and stiff stability. Additional appendices discuss linear multistep methods, discuss directions for further work, and exhibit numerical analysis results for some other commonly used lower-order methods.
Large linear magnetoresistance from neutral defects in Bi$_2$Se$_3$
Kumar, Devendra; Lakhani, Archana
2016-01-01
The chalcogenide Bi$_2$Se$_3$ can attain the three dimensional (3D) Dirac semimetal state under the influence of strain and microstrain. Here we report the presnece of large linear magnetoresistance in such a Bi$_2$Se$_3$ crystal. The magnetoresistance has quadratic form at low fields which crossovers to linear above 4 T. The temperature dependence of magnetoresistance scales with carrier mobility and the crossover field scales with inverse of mobility. Our analysis suggest that the linear ma...
Large-time asymptotic behaviour of solutions of non-linear Sobolev-type equations
International Nuclear Information System (INIS)
Kaikina, Elena I; Naumkin, Pavel I; Shishmarev, Il'ya A
2009-01-01
The large-time asymptotic behaviour of solutions of the Cauchy problem is investigated for a non-linear Sobolev-type equation with dissipation. For small initial data the approach taken is based on a detailed analysis of the Green's function of the linear problem and the use of the contraction mapping method. The case of large initial data is also closely considered. In the supercritical case the asymptotic formulae are quasi-linear. The asymptotic behaviour of solutions of a non-linear Sobolev-type equation with a critical non-linearity of the non-convective kind differs by a logarithmic correction term from the behaviour of solutions of the corresponding linear equation. For a critical convective non-linearity, as well as for a subcritical non-convective non-linearity it is proved that the leading term of the asymptotic expression for large times is a self-similar solution. For Sobolev equations with convective non-linearity the asymptotic behaviour of solutions in the subcritical case is the product of a rarefaction wave and a shock wave. Bibliography: 84 titles.
Infeasible Interior-Point Methods for Linear Optimization Based on Large Neighborhood
Asadi, A.R.; Roos, C.
2015-01-01
In this paper, we design a class of infeasible interior-point methods for linear optimization based on large neighborhood. The algorithm is inspired by a full-Newton step infeasible algorithm with a linear convergence rate in problem dimension that was recently proposed by the second author.
Solving Man-Induced Large-Scale Conservation Problems: The Spanish Imperial Eagle and Power Lines
López-López, Pascual; Ferrer, Miguel; Madero, Agustín; Casado, Eva; McGrady, Michael
2011-01-01
Background Man-induced mortality of birds caused by electrocution with poorly-designed pylons and power lines has been reported to be an important mortality factor that could become a major cause of population decline of one of the world rarest raptors, the Spanish imperial eagle (Aquila adalberti). Consequently it has resulted in an increasing awareness of this problem amongst land managers and the public at large, as well as increased research into the distribution of electrocution events and likely mitigation measures. Methodology/Principal Findings We provide information of how mitigation measures implemented on a regional level under the conservation program of the Spanish imperial eagle have resulted in a positive shift of demographic trends in Spain. A 35 years temporal data set (1974–2009) on mortality of Spanish imperial eagle was recorded, including population censuses, and data on electrocution and non-electrocution of birds. Additional information was obtained from 32 radio-tracked young eagles and specific field surveys. Data were divided into two periods, before and after the approval of a regional regulation of power line design in 1990 which established mandatory rules aimed at minimizing or eliminating the negative impacts of power lines facilities on avian populations. Our results show how population size and the average annual percentage of population change have increased between the two periods, whereas the number of electrocuted birds has been reduced in spite of the continuous growing of the wiring network. Conclusions Our results demonstrate that solving bird electrocution is an affordable problem if political interest is shown and financial investment is made. The combination of an adequate spatial planning with a sustainable development of human infrastructures will contribute positively to the conservation of the Spanish imperial eagle and may underpin population growth and range expansion, with positive side effects on other endangered
International Nuclear Information System (INIS)
Chiche, A.
2012-01-01
This manuscript deals with large-scale optimization problems, and more specifically with solving the electricity unit commitment problem arising at EDF. First, we focused on the augmented Lagrangian algorithm. The behavior of that algorithm on an infeasible convex quadratic optimization problem is analyzed. It is shown that the algorithm finds a point that satisfies the shifted constraints with the smallest possible shift in the sense of the Euclidean norm and that it minimizes the objective on the corresponding shifted constrained set. The convergence to such a point is realized at a global linear rate, which depends explicitly on the augmentation parameter. This suggests us a rule for determining the augmentation parameter to control the speed of convergence of the shifted constraint norm to zero. This rule has the advantage of generating bounded augmentation parameters even when the problem is infeasible. As a by-product, the algorithm computes the smallest translation in the Euclidean norm that makes the constraints feasible. Furthermore, this work provides solution methods for stochastic optimization industrial problems decomposed on a scenario tree, based on the progressive hedging algorithm introduced by [Rockafellar et Wets, 1991]. We also focus on the convergence of that algorithm. On the one hand, we offer a counter-example showing that the algorithm could diverge if its augmentation parameter is iteratively updated. On the other hand, we show how to recover the multipliers associated with the non-dualized constraints defined on the scenario tree from those associated with the corresponding constraints of the scenario subproblems. Their convergence is also analyzed for convex problems. The practical interest of theses solutions techniques is corroborated by numerical experiments performed on the electric production management problem. We apply the progressive hedging algorithm to a realistic industrial problem. More precisely, we solve the French medium
Large linear magnetoresistance and magnetothermopower in layered SrZnSb$_2$
Wang, Kefeng; Petrovic, C.
2016-01-01
We report the large linear magnetoresistance ($\\sim 300\\%$ in 9 T field at 2 K) and magnetothermopower in layered SrZnSb$_2$ crystal with quasi-two-dimensional Sb layers. A crossover from the semiclassical parabolic field dependent magnetoresistance to linear field dependent magnetoresistance with increasing magnetic field is observed. The magnetoresistance behavior can be described very well by combining the semiclassical cyclotron contribution and the quantum limit magnetoresistance. Magnet...
Solving large test-day models by iteration on data and preconditioned conjugate gradient.
Lidauer, M; Strandén, I; Mäntysaari, E A; Pösö, J; Kettunen, A
1999-12-01
A preconditioned conjugate gradient method was implemented into an iteration on a program for data estimation of breeding values, and its convergence characteristics were studied. An algorithm was used as a reference in which one fixed effect was solved by Gauss-Seidel method, and other effects were solved by a second-order Jacobi method. Implementation of the preconditioned conjugate gradient required storing four vectors (size equal to number of unknowns in the mixed model equations) in random access memory and reading the data at each round of iteration. The preconditioner comprised diagonal blocks of the coefficient matrix. Comparison of algorithms was based on solutions of mixed model equations obtained by a single-trait animal model and a single-trait, random regression test-day model. Data sets for both models used milk yield records of primiparous Finnish dairy cows. Animal model data comprised 665,629 lactation milk yields and random regression test-day model data of 6,732,765 test-day milk yields. Both models included pedigree information of 1,099,622 animals. The animal model ¿random regression test-day model¿ required 122 ¿305¿ rounds of iteration to converge with the reference algorithm, but only 88 ¿149¿ were required with the preconditioned conjugate gradient. To solve the random regression test-day model with the preconditioned conjugate gradient required 237 megabytes of random access memory and took 14% of the computation time needed by the reference algorithm.
Ebrahimnejad, Ali
2015-08-01
There are several methods, in the literature, for solving fuzzy variable linear programming problems (fuzzy linear programming in which the right-hand-side vectors and decision variables are represented by trapezoidal fuzzy numbers). In this paper, the shortcomings of some existing methods are pointed out and to overcome these shortcomings a new method based on the bounded dual simplex method is proposed to determine the fuzzy optimal solution of that kind of fuzzy variable linear programming problems in which some or all variables are restricted to lie within lower and upper bounds. To illustrate the proposed method, an application example is solved and the obtained results are given. The advantages of the proposed method over existing methods are discussed. Also, one application of this algorithm in solving bounded transportation problems with fuzzy supplies and demands is dealt with. The proposed method is easy to understand and to apply for determining the fuzzy optimal solution of bounded fuzzy variable linear programming problems occurring in real-life situations.
Large linear magnetoresistance in topological crystalline insulator Pb_0_._6Sn_0_._4Te
International Nuclear Information System (INIS)
Roychowdhury, Subhajit; Ghara, Somnath; Guin, Satya N.; Sundaresan, A.; Biswas, Kanishka
2016-01-01
Classical magnetoresistance generally follows the quadratic dependence of the magnetic field at lower field and finally saturates when field is larger. Here, we report the large positive non-saturating linear magnetoresistance in topological crystalline insulator, Pb_0_._6Sn_0_._4Te, at different temperatures between 3 K and 300 K in magnetic field up to 9 T. Magnetoresistance value as high as ∼200% was achieved at 3 K at magnetic field of 9 T. Linear magnetoresistance observed in Pb_0_._6Sn_0_._4Te is mainly governed by the spatial fluctuation carrier mobility due to distortions in the current paths in inhomogeneous conductor. - Graphical abstract: Large non-saturating linear magnetoresistance has been evidenced in topological crystalline insulator, Pb_0_._6Sn_0_._4Te, at different temperatures between 3 K and 300 K in magnetic field up to 9 T. - Highlights: • Large non-saturating linear magnetoresistance was achieved in the topological crystalline insulator, Pb_0_._6Sn_0_._4Te. • Highest magnetoresistance value as high as ~200% was achieved at 3 K at magnetic field of 9 T. • Linear magnetoresistance in Pb_0_._6Sn_0_._4Te is mainly governed by the spatial fluctuation of the carrier mobility.
Large Negative Linear Compressibility in InH(BDC)₂ from Framework Hinging.
Zeng, Qingxin; Wang, Kai; Zou, Bo
2017-11-08
Materials with negative linear compressibility (NLC) counterintuitively expand along one specific direction coupled to the volume reduction when compressed uniformly. NLC with a large value is desired for compression and materials science. However, NLC is generally smaller than -20 TPa -1 . High-pressure X-ray diffraction experiments reveal that the β-quartz-like InH(BDC) 2 generates an extreme NLC (-62.4 TPa -1 ) by framework hinging. InH(BDC) 2 is much safer and lower-cost than Au + /Ag + and CN - -containing materials that dominated the fields of large NLC. This work reconfirms that a negative thermal expansion flexible framework could likely exhibit large NLC. Moreover, a large NLC could be anticipated to arise from β-quartz-like or related frameworks composed of rigid linear ligands and flexible framework angles.
Seismic evaluation of a large nuclear pump bearing using non-linear dynamic analysis
International Nuclear Information System (INIS)
Huber, K.A.; Hugins, M.S.
1983-01-01
Hydrostatic bearings of a large vertical pump using sodium as the lubricant were critically examined to determine their ability to withstand seismic loads. Initial linear dynamics analyses predicted journal displacements to exceed bearing clearance by a ratio of 3:1. Equivalent time-history excitations were then developed from the response spectra to determine the number, magnitude, and duration of the bearing impact loads. Predicted loads were further reduced by 50% by modeling non-linear bearing characteristics normally present but not generally included in conventional linear analyses. Results are presented of the comprehensive design evaluation performed, based on these non-linear predictions, that assess stress, wear, and fatigue to demonstrate hydrostatic bearing integrity
Robust estimation for partially linear models with large-dimensional covariates.
Zhu, LiPing; Li, RunZe; Cui, HengJian
2013-10-01
We are concerned with robust estimation procedures to estimate the parameters in partially linear models with large-dimensional covariates. To enhance the interpretability, we suggest implementing a noncon-cave regularization method in the robust estimation procedure to select important covariates from the linear component. We establish the consistency for both the linear and the nonlinear components when the covariate dimension diverges at the rate of [Formula: see text], where n is the sample size. We show that the robust estimate of linear component performs asymptotically as well as its oracle counterpart which assumes the baseline function and the unimportant covariates were known a priori. With a consistent estimator of the linear component, we estimate the nonparametric component by a robust local linear regression. It is proved that the robust estimate of nonlinear component performs asymptotically as well as if the linear component were known in advance. Comprehensive simulation studies are carried out and an application is presented to examine the finite-sample performance of the proposed procedures.
A Decomposition-Based Pricing Method for Solving a Large-Scale MILP Model for an Integrated Fishery
Directory of Open Access Journals (Sweden)
M. Babul Hasan
2007-01-01
The IFP can be decomposed into a trawler-scheduling subproblem and a fish-processing subproblem in two different ways by relaxing different sets of constraints. We tried conventional decomposition techniques including subgradient optimization and Dantzig-Wolfe decomposition, both of which were unacceptably slow. We then developed a decomposition-based pricing method for solving the large fishery model, which gives excellent computation times. Numerical results for several planning horizon models are presented.
Non-linear finite element analyses applicable for the design of large reinforced concrete structures
Engen, M; Hendriks, M.A.N.; Øverli, Jan Arve; Åldstedt, Erik
2017-01-01
In order to make non-linear finite element analyses applicable during assessments of the ultimate load capacity or the structural reliability of large reinforced concrete structures, there is need for an efficient solution strategy with a low modelling uncertainty. A solution strategy comprises
Material model for non-linear finite element analyses of large concrete structures
Engen, Morten; Hendriks, M.A.N.; Øverli, Jan Arve; Åldstedt, Erik; Beushausen, H.
2016-01-01
A fully triaxial material model for concrete was implemented in a commercial finite element code. The only required input parameter was the cylinder compressive strength. The material model was suitable for non-linear finite element analyses of large concrete structures. The importance of including
Modeling containment of large wildfires using generalized linear mixed-model analysis
Mark Finney; Isaac C. Grenfell; Charles W. McHugh
2009-01-01
Billions of dollars are spent annually in the United States to contain large wildland fires, but the factors contributing to suppression success remain poorly understood. We used a regression model (generalized linear mixed-model) to model containment probability of individual fires, assuming that containment was a repeated-measures problem (fixed effect) and...
Use of personal computers in performing a linear modal analysis of a large finite-element model
International Nuclear Information System (INIS)
Wagenblast, G.R.
1991-01-01
This paper presents the use of personal computers in performing a dynamic frequency analysis of a large (2,801 degrees of freedom) finite-element model. Large model linear time history dynamic evaluations of safety related structures were previously restricted to mainframe computers using direct integration analysis methods. This restriction was a result of the limited memory and speed of personal computers. With the advances in memory capacity and speed of the personal computers, large finite-element problems now can be solved in the office in a timely and cost effective manner. Presented in three sections, this paper describes the procedure used to perform the dynamic frequency analysis of the large (2,801 degrees of freedom) finite-element model on a personal computer. Section 2.0 describes the structure and the finite-element model that was developed to represent the structure for use in the dynamic evaluation. Section 3.0 addresses the hardware and software used to perform the evaluation and the optimization of the hardware and software operating configuration to minimize the time required to perform the analysis. Section 4.0 explains the analysis techniques used to reduce the problem to a size compatible with the hardware and software memory capacity and configuration
Sole, Marla A.
2016-01-01
Open-ended questions that can be solved using different strategies help students learn and integrate content, and provide teachers with greater insights into students' unique capabilities and levels of understanding. This article provides a problem that was modified to allow for multiple approaches. Students tended to employ high-powered, complex,…
DEFF Research Database (Denmark)
Bendtsen, Claus; Nielsen, Ole Holm; Hansen, Lars Bruno
2001-01-01
The quantum mechanical ground state of electrons is described by Density Functional Theory, which leads to large minimization problems. An efficient minimization method uses a self-consistent field (SCF) solution of large eigenvalue problems. The iterative Davidson algorithm is often used, and we...
Bui-Thanh, T.; Girolami, M.
2014-11-01
We consider the Riemann manifold Hamiltonian Monte Carlo (RMHMC) method for solving statistical inverse problems governed by partial differential equations (PDEs). The Bayesian framework is employed to cast the inverse problem into the task of statistical inference whose solution is the posterior distribution in infinite dimensional parameter space conditional upon observation data and Gaussian prior measure. We discretize both the likelihood and the prior using the H1-conforming finite element method together with a matrix transfer technique. The power of the RMHMC method is that it exploits the geometric structure induced by the PDE constraints of the underlying inverse problem. Consequently, each RMHMC posterior sample is almost uncorrelated/independent from the others providing statistically efficient Markov chain simulation. However this statistical efficiency comes at a computational cost. This motivates us to consider computationally more efficient strategies for RMHMC. At the heart of our construction is the fact that for Gaussian error structures the Fisher information matrix coincides with the Gauss-Newton Hessian. We exploit this fact in considering a computationally simplified RMHMC method combining state-of-the-art adjoint techniques and the superiority of the RMHMC method. Specifically, we first form the Gauss-Newton Hessian at the maximum a posteriori point and then use it as a fixed constant metric tensor throughout RMHMC simulation. This eliminates the need for the computationally costly differential geometric Christoffel symbols, which in turn greatly reduces computational effort at a corresponding loss of sampling efficiency. We further reduce the cost of forming the Fisher information matrix by using a low rank approximation via a randomized singular value decomposition technique. This is efficient since a small number of Hessian-vector products are required. The Hessian-vector product in turn requires only two extra PDE solves using the adjoint
Modeling and solving a large-scale generation expansion planning problem under uncertainty
Energy Technology Data Exchange (ETDEWEB)
Jin, Shan; Ryan, Sarah M. [Iowa State University, Department of Industrial and Manufacturing Systems Engineering, Ames (United States); Watson, Jean-Paul [Sandia National Laboratories, Discrete Math and Complex Systems Department, Albuquerque (United States); Woodruff, David L. [University of California Davis, Graduate School of Management, Davis (United States)
2011-11-15
We formulate a generation expansion planning problem to determine the type and quantity of power plants to be constructed over each year of an extended planning horizon, considering uncertainty regarding future demand and fuel prices. Our model is expressed as a two-stage stochastic mixed-integer program, which we use to compute solutions independently minimizing the expected cost and the Conditional Value-at-Risk; i.e., the risk of significantly larger-than-expected operational costs. We introduce stochastic process models to capture demand and fuel price uncertainty, which are in turn used to generate trees that accurately represent the uncertainty space. Using a realistic problem instance based on the Midwest US, we explore two fundamental, unexplored issues that arise when solving any stochastic generation expansion model. First, we introduce and discuss the use of an algorithm for computing confidence intervals on obtained solution costs, to account for the fact that a finite sample of scenarios was used to obtain a particular solution. Second, we analyze the nature of solutions obtained under different parameterizations of this method, to assess whether the recommended solutions themselves are invariant to changes in costs. The issues are critical for decision makers who seek truly robust recommendations for generation expansion planning. (orig.)
Energy Technology Data Exchange (ETDEWEB)
Nygaard, K
1968-09-15
From the point of view that no mathematical method can ever minimise or alter errors already made in a physical measurement, the classical least squares method has severe limitations which makes it unsuitable for the statistical analysis of many physical measurements. Based on the assumptions that the experimental errors are characteristic for each single experiment and that the errors must be properly estimated rather than minimised, a new method for solving large systems of linear equations is developed. The new method exposes the entire range of possible solutions before the decision is taken which of the possible solutions should be chosen as a representative one. The choice is based on physical considerations which (in two examples, curve fitting and unfolding of a spectrum) are presented in such a form that a computer is able to make the decision, A description of the computation is given. The method described is a tool for removing uncertainties due to conventional mathematical formulations (zero determinant, linear dependence) and which are not inherent in the physical problem as such. The method is therefore especially well fitted for unfolding of spectra.
International Nuclear Information System (INIS)
Nygaard, K.
1968-09-01
From the point of view that no mathematical method can ever minimise or alter errors already made in a physical measurement, the classical least squares method has severe limitations which makes it unsuitable for the statistical analysis of many physical measurements. Based on the assumptions that the experimental errors are characteristic for each single experiment and that the errors must be properly estimated rather than minimised, a new method for solving large systems of linear equations is developed. The new method exposes the entire range of possible solutions before the decision is taken which of the possible solutions should be chosen as a representative one. The choice is based on physical considerations which (in two examples, curve fitting and unfolding of a spectrum) are presented in such a form that a computer is able to make the decision, A description of the computation is given. The method described is a tool for removing uncertainties due to conventional mathematical formulations (zero determinant, linear dependence) and which are not inherent in the physical problem as such. The method is therefore especially well fitted for unfolding of spectra
Pang, Yu; Zhang, Kunning; Yang, Zhen; Jiang, Song; Ju, Zhenyi; Li, Yuxing; Wang, Xuefeng; Wang, Danyang; Jian, Muqiang; Zhang, Yingying; Liang, Renrong; Tian, He; Yang, Yi; Ren, Tian-Ling
2018-03-27
Recently, wearable pressure sensors have attracted tremendous attention because of their potential applications in monitoring physiological signals for human healthcare. Sensitivity and linearity are the two most essential parameters for pressure sensors. Although various designed micro/nanostructure morphologies have been introduced, the trade-off between sensitivity and linearity has not been well balanced. Human skin, which contains force receptors in a reticular layer, has a high sensitivity even for large external stimuli. Herein, inspired by the skin epidermis with high-performance force sensing, we have proposed a special surface morphology with spinosum microstructure of random distribution via the combination of an abrasive paper template and reduced graphene oxide. The sensitivity of the graphene pressure sensor with random distribution spinosum (RDS) microstructure is as high as 25.1 kPa -1 in a wide linearity range of 0-2.6 kPa. Our pressure sensor exhibits superior comprehensive properties compared with previous surface-modified pressure sensors. According to simulation and mechanism analyses, the spinosum microstructure and random distribution contribute to the high sensitivity and large linearity range, respectively. In addition, the pressure sensor shows promising potential in detecting human physiological signals, such as heartbeat, respiration, phonation, and human motions of a pushup, arm bending, and walking. The wearable pressure sensor array was further used to detect gait states of supination, neutral, and pronation. The RDS microstructure provides an alternative strategy to improve the performance of pressure sensors and extend their potential applications in monitoring human activities.
Three-point phase correlations: A new measure of non-linear large-scale structure
Wolstenhulme, Richard; Obreschkow, Danail
2015-01-01
We derive an analytical expression for a novel large-scale structure observable: the line correlation function. The line correlation function, which is constructed from the three-point correlation function of the phase of the density field, is a robust statistical measure allowing the extraction of information in the non-linear and non-Gaussian regime. We show that, in perturbation theory, the line correlation is sensitive to the coupling kernel F_2, which governs the non-linear gravitational evolution of the density field. We compare our analytical expression with results from numerical simulations and find a very good agreement for separations r>20 Mpc/h. Fitting formulae for the power spectrum and the non-linear coupling kernel at small scales allow us to extend our prediction into the strongly non-linear regime. We discuss the advantages of the line correlation relative to standard statistical measures like the bispectrum. Unlike the latter, the line correlation is independent of the linear bias. Furtherm...
International Nuclear Information System (INIS)
Soltani, J.; Fath Abadi, A.M.
2003-01-01
This paper describes the application of static var compensators, on an electrical distribution network containing two large synchronous motors, one of which is excited via a three-phase thyristor bridge rectifier. The second machine is excited via a diode bridge rectifier. Based on linear optimization control, the measurable feedback signals are applied to the control system loops of static var compensators and the excitation control loop of the first synchronous motor. The phase equations method was used to develop a computer program to model the distribution network. Computer results were obtained to demonstrate the system performance for some abnormal modes of operation. These results show that employing static var compensators based on the linear optimization control design for electrical distribution networks containing large synchronous motors is beneficial and may be considered a first stage of the system design
Leon, Stéphane; Bergond, Gilles; Vallenari, Antonella
1999-04-01
We present the tidal tail distributions of a sample of candidate binary clusters located in the bar of the Large Magellanic Cloud (LMC). One isolated cluster, SL 268, is presented in order to study the effect of the LMC tidal field. All the candidate binary clusters show tidal tails, confirming that the pairs are formed by physically linked objects. The stellar mass in the tails covers a large range, from 1.8x 10(3) to 3x 10(4) \\msun. We derive a total mass estimate for SL 268 and SL 356. At large radii, the projected density profiles of SL 268 and SL 356 fall off as r(-gamma ) , with gamma = 2.27 and gamma =3.44, respectively. Out of 4 pairs or multiple systems, 2 are older than the theoretical survival time of binary clusters (going from a few 10(6) years to 10(8) years). A pair shows too large age difference between the components to be consistent with classical theoretical models of binary cluster formation (Fujimoto & Kumai \\cite{fujimoto97}). We refer to this as the ``overmerging'' problem. A different scenario is proposed: the formation proceeds in large molecular complexes giving birth to groups of clusters over a few 10(7) years. In these groups the expected cluster encounter rate is larger, and tidal capture has higher probability. Cluster pairs are not born together through the splitting of the parent cloud, but formed later by tidal capture. For 3 pairs, we tentatively identify the star cluster group (SCG) memberships. The SCG formation, through the recent cluster starburst triggered by the LMC-SMC encounter, in contrast with the quiescent open cluster formation in the Milky Way can be an explanation to the paucity of binary clusters observed in our Galaxy. Based on observations collected at the European Southern Observatory, La Silla, Chile}
Evaluation of linear DC motor actuators for control of large space structures
Ide, Eric Nelson
1988-01-01
This thesis examines the use of a linear DC motor as a proof mass actuator for the control of large space structures. A model for the actuator, including the current and force compensation used, is derived. Because of the force compensation, the actuator is unstable when placed on a structure. Relative position feedback is used for actuator stabilization. This method of compensation couples the actuator to the mast in a feedback configuration. Three compensator designs are prop...
Klegeris, Andis; Hurren, Heather
2011-12-01
Problem-based learning (PBL) can be described as a learning environment where the problem drives the learning. This technique usually involves learning in small groups, which are supervised by tutors. It is becoming evident that PBL in a small-group setting has a robust positive effect on student learning and skills, including better problem-solving skills and an increase in overall motivation. However, very little research has been done on the educational benefits of PBL in a large classroom setting. Here, we describe a PBL approach (using tutorless groups) that was introduced as a supplement to standard didactic lectures in University of British Columbia Okanagan undergraduate biochemistry classes consisting of 45-85 students. PBL was chosen as an effective method to assist students in learning biochemical and physiological processes. By monitoring student attendance and using informal and formal surveys, we demonstrated that PBL has a significant positive impact on student motivation to attend and participate in the course work. Student responses indicated that PBL is superior to traditional lecture format with regard to the understanding of course content and retention of information. We also demonstrated that student problem-solving skills are significantly improved, but additional controlled studies are needed to determine how much PBL exercises contribute to this improvement. These preliminary data indicated several positive outcomes of using PBL in a large classroom setting, although further studies aimed at assessing student learning are needed to further justify implementation of this technique in courses delivered to large undergraduate classes.
International Nuclear Information System (INIS)
Aspinall, J.
1982-01-01
A computational method was developed which alleviates the need for lengthy parametric scans as part of a design process. The method makes use of a least squares algorithm to find the optimal value of a parameter vector. Optimal is defined in terms of a utility function prescribed by the user. The placement of the vertical field coils of a torsatron is such a non linear problem
International Nuclear Information System (INIS)
Ahlfeld, D.P.; Dougherty, D.E.
1994-11-01
MODLP is a computational tool that may help design capture zones for controlling the movement of contaminated groundwater. It creates and solves linear optimization programs that contain constraints on hydraulic head or head differences in a groundwater system. The groundwater domain is represented by USGS MODFLOW groundwater flow simulation model. This document describes the general structure of the computer program, MODLP, the types of constraints that may be imposed, detailed input instructions, interpretation of the output, and the interaction with the MODFLOW simulation kernel
Several Families of Sequences with Low Correlation and Large Linear Span
Zeng, Fanxin; Zhang, Zhenyu
In DS-CDMA systems and DS-UWB radios, low correlation of spreading sequences can greatly help to minimize multiple access interference (MAI) and large linear span of spreading sequences can reduce their predictability. In this letter, new sequence sets with low correlation and large linear span are proposed. Based on the construction Trm1[Trnm(αbt+γiαdt)]r for generating p-ary sequences of period pn-1, where n=2m, d=upm±v, b=u±v, γi∈GF(pn), and p is an arbitrary prime number, several methods to choose the parameter d are provided. The obtained sequences with family size pn are of four-valued, five-valued, six-valued or seven-valued correlation and the maximum nontrivial correlation value is (u+v-1)pm-1. The simulation by a computer shows that the linear span of the new sequences is larger than that of the sequences with Niho-type and Welch-type decimations, and similar to that of [10].
Frequency-scanning MALDI linear ion trap mass spectrometer for large biomolecular ion detection.
Lu, I-Chung; Lin, Jung Lee; Lai, Szu-Hsueh; Chen, Chung-Hsuan
2011-11-01
This study presents the first report on the development of a matrix-assisted laser desorption ionization (MALDI) linear ion trap mass spectrometer for large biomolecular ion detection by frequency scan. We designed, installed, and tested this radio frequency (RF) scan linear ion trap mass spectrometer and its associated electronics to dramatically extend the mass region to be detected. The RF circuit can be adjusted from 300 to 10 kHz with a set of operation amplifiers. To trap the ions produced by MALDI, a high pressure of helium buffer gas was employed to quench extra kinetic energy of the heavy ions produced by MALDI. The successful detection of the singly charged secretory immunoglobulin A ions indicates that the detectable mass-to-charge ratio (m/z) of this system can reach ~385 000 or beyond.
DEFF Research Database (Denmark)
Stolpe, Mathias; Bendsøe, Martin P.
2007-01-01
This paper present some initial results pertaining to a search for globally optimal solutions to a challenging benchmark example proposed by Zhou and Rozvany. This means that we are dealing with global optimization of the classical single load minimum compliance topology design problem with a fixed...... finite element discretization and with discrete design variables. Global optimality is achieved by the implementation of some specially constructed convergent nonlinear branch and cut methods, based on the use of natural relaxations and by applying strengthening constraints (linear valid inequalities...
DEFF Research Database (Denmark)
Stolpe, Mathias; Bendsøe, Martin P.
2007-01-01
This paper present some initial results pertaining to a search for globally optimal solutions to a challenging benchmark example proposed by Zhou and Rozvany. This means that we are dealing with global optimization of the classical single load minimum compliance topology design problem with a fixed...... finite element discretization and with discrete design variables. Global optimality is achieved by the implementation of some specially constructed convergent nonlinear branch and cut methods, based on the use of natural relaxations and by applying strengthening constraints (linear valid inequalities......) and cuts....
Minimization of Linear Functionals Defined on| Solutions of Large-Scale Discrete Ill-Posed Problems
DEFF Research Database (Denmark)
Elden, Lars; Hansen, Per Christian; Rojas, Marielba
2003-01-01
The minimization of linear functionals de ned on the solutions of discrete ill-posed problems arises, e.g., in the computation of con dence intervals for these solutions. In 1990, Elden proposed an algorithm for this minimization problem based on a parametric-programming reformulation involving...... the solution of a sequence of trust-region problems, and using matrix factorizations. In this paper, we describe MLFIP, a large-scale version of this algorithm where a limited-memory trust-region solver is used on the subproblems. We illustrate the use of our algorithm in connection with an inverse heat...
Pseudoinverse preconditioners and iterative methods for large dense linear least-squares problems
Directory of Open Access Journals (Sweden)
Oskar Cahueñas
2013-05-01
Full Text Available We address the issue of approximating the pseudoinverse of the coefficient matrix for dynamically building preconditioning strategies for the numerical solution of large dense linear least-squares problems. The new preconditioning strategies are embedded into simple and well-known iterative schemes that avoid the use of the, usually ill-conditioned, normal equations. We analyze a scheme to approximate the pseudoinverse, based on Schulz iterative method, and also different iterative schemes, based on extensions of Richardson's method, and the conjugate gradient method, that are suitable for preconditioning strategies. We present preliminary numerical results to illustrate the advantages of the proposed schemes.
Energy Technology Data Exchange (ETDEWEB)
Tetsu, Hiroyuki; Nakamoto, Taishi, E-mail: h.tetsu@geo.titech.ac.jp [Earth and Planetary Sciences, Tokyo Institute of Technology, Tokyo 152-8551 (Japan)
2016-03-15
Radiation is an important process of energy transport, a force, and a basis for synthetic observations, so radiation hydrodynamics (RHD) calculations have occupied an important place in astrophysics. However, although the progress in computational technology is remarkable, their high numerical cost is still a persistent problem. In this work, we compare the following schemes used to solve the nonlinear simultaneous equations of an RHD algorithm with the flux-limited diffusion approximation: the Newton–Raphson (NR) method, operator splitting, and linearization (LIN), from the perspective of the computational cost involved. For operator splitting, in addition to the traditional simple operator splitting (SOS) scheme, we examined the scheme developed by Douglas and Rachford (DROS). We solve three test problems (the thermal relaxation mode, the relaxation and the propagation of linear waves, and radiating shock) using these schemes and then compare their dependence on the time step size. As a result, we find the conditions of the time step size necessary for adopting each scheme. The LIN scheme is superior to other schemes if the ratio of radiation pressure to gas pressure is sufficiently low. On the other hand, DROS can be the most efficient scheme if the ratio is high. Although the NR scheme can be adopted independently of the regime, especially in a problem that involves optically thin regions, the convergence tends to be worse. In all cases, SOS is not practical.
Herman, Gabor T; Chen, Wei
2008-03-01
The goal of Intensity-Modulated Radiation Therapy (IMRT) is to deliver sufficient doses to tumors to kill them, but without causing irreparable damage to critical organs. This requirement can be formulated as a linear feasibility problem. The sequential (i.e., iteratively treating the constraints one after another in a cyclic fashion) algorithm ART3 is known to find a solution to such problems in a finite number of steps, provided that the feasible region is full dimensional. We present a faster algorithm called ART3+. The idea of ART3+ is to avoid unnecessary checks on constraints that are likely to be satisfied. The superior performance of the new algorithm is demonstrated by mathematical experiments inspired by the IMRT application.
International Nuclear Information System (INIS)
Kanamori, Takahiro; Kamata, Shouji; Ito, Shinichi.
1989-01-01
A prototype high energy X-ray CT (computed tomography) system has been developed which employs a linear accelerator as the X-ray source (max. photon energy: 12 MeV). One problem encountered in development of this CT system was to reduce the scattered photons from adjacent detectors, i.e. crosstalk, due to high energy X-rays. This crosstalk was reduced to 2% by means of detector shields using tungsten spacers. Spatial resolution was not affected by such small crosstalk as confirmed by numerical simulations. A second problem was to reduce the scattered photons from the test object. This was done using collimators. A third concern was to realize a wide dynamic range data processing which would allow applications to large and dense objects. This problem was solved by using a sample and hold data acquisition method to reduce the dark current of the photo detectors. The dynamic range of this system was experimentally confirmed over 60 dB. It was demonstrated that slits (width: 2 mm) in an iron object (diameter: 25 cm) could be imaged by this prototype CT system. (author)
Liu, Tianyang; Chan, Hiu Ning; Grimshaw, Roger; Chow, Kwok Wing
2017-11-01
The spatial structure of small disturbances in stratified flows without background shear, usually named the `Taylor-Goldstein equation', is studied by employing the Boussinesq approximation (variation in density ignored except in the buoyancy). Analytical solutions are derived for special wavenumbers when the Brunt-Väisälä frequency is quadratic in hyperbolic secant, by comparison with coupled systems of nonlinear Schrödinger equations intensively studied in the literature. Cases of coupled Schrödinger equations with four, five and six components are utilized as concrete examples. Dispersion curves for arbitrary wavenumbers are obtained numerically. The computations of the group velocity, second harmonic, induced mean flow, and the second derivative of the angular frequency can all be facilitated by these exact linear eigenfunctions of the Taylor-Goldstein equation in terms of hyperbolic function, leading to a cubic Schrödinger equation for the evolution of a wavepacket. The occurrence of internal rogue waves can be predicted if the dispersion and cubic nonlinearity terms of the Schrödinger equations are of the same sign. Partial financial support has been provided by the Research Grants Council contract HKU 17200815.
Energy Technology Data Exchange (ETDEWEB)
Alleon, G. [EADS-CCR, 31 - Blagnac (France); Carpentieri, B.; Du, I.S.; Giraud, L.; Langou, J.; Martin, E. [Cerfacs, 31 - Toulouse (France)
2003-07-01
The boundary element method has become a popular tool for the solution of Maxwell's equations in electromagnetism. It discretizes only the surface of the radiating object and gives rise to linear systems that are smaller in size compared to those arising from finite element or finite difference discretizations. However, these systems are prohibitively demanding in terms of memory for direct methods and challenging to solve by iterative methods. In this paper we address the iterative solution via preconditioned Krylov methods of electromagnetic scattering problems expressed in an integral formulation, with main focus on the design of the pre-conditioner. We consider an approximate inverse method based on the Frobenius-norm minimization with a pattern prescribed in advance. The pre-conditioner is constructed from a sparse approximation of the dense coefficient matrix, and the patterns both for the pre-conditioner and for the coefficient matrix are computed a priori using geometric information from the mesh. We describe the implementation of the approximate inverse in an out-of-core parallel code that uses multipole techniques for the matrix-vector products, and show results on the numerical scalability of our method on systems of size up to one million unknowns. We propose an embedded iterative scheme based on the GMRES method and combined with multipole techniques, aimed at improving the robustness of the approximate inverse for large problems. We prove by numerical experiments that the proposed scheme enables the solution of very large and difficult problems efficiently at reduced computational and memory cost. Finally we perform a preliminary study on a spectral two-level pre-conditioner to enhance the robustness of our method. This numerical technique exploits spectral information of the preconditioned systems to build a low rank-update of the pre-conditioner. (authors)
International Nuclear Information System (INIS)
Alleon, G.; Carpentieri, B.; Du, I.S.; Giraud, L.; Langou, J.; Martin, E.
2003-01-01
The boundary element method has become a popular tool for the solution of Maxwell's equations in electromagnetism. It discretizes only the surface of the radiating object and gives rise to linear systems that are smaller in size compared to those arising from finite element or finite difference discretizations. However, these systems are prohibitively demanding in terms of memory for direct methods and challenging to solve by iterative methods. In this paper we address the iterative solution via preconditioned Krylov methods of electromagnetic scattering problems expressed in an integral formulation, with main focus on the design of the pre-conditioner. We consider an approximate inverse method based on the Frobenius-norm minimization with a pattern prescribed in advance. The pre-conditioner is constructed from a sparse approximation of the dense coefficient matrix, and the patterns both for the pre-conditioner and for the coefficient matrix are computed a priori using geometric information from the mesh. We describe the implementation of the approximate inverse in an out-of-core parallel code that uses multipole techniques for the matrix-vector products, and show results on the numerical scalability of our method on systems of size up to one million unknowns. We propose an embedded iterative scheme based on the GMRES method and combined with multipole techniques, aimed at improving the robustness of the approximate inverse for large problems. We prove by numerical experiments that the proposed scheme enables the solution of very large and difficult problems efficiently at reduced computational and memory cost. Finally we perform a preliminary study on a spectral two-level pre-conditioner to enhance the robustness of our method. This numerical technique exploits spectral information of the preconditioned systems to build a low rank-update of the pre-conditioner. (authors)
Large angle and high linearity two-dimensional laser scanner based on voice coil actuators
Wu, Xin; Chen, Sihai; Chen, Wei; Yang, Minghui; Fu, Wen
2011-10-01
A large angle and high linearity two-dimensional laser scanner with an in-house ingenious deflection angle detecting system is developed based on voice coil actuators direct driving mechanism. The specially designed voice coil actuators make the steering mirror moving at a sufficiently large angle. Frequency sweep method based on virtual instruments is employed to achieve the natural frequency of the laser scanner. The response shows that the performance of the laser scanner is limited by the mechanical resonances. The closed-loop controller based on mathematical model is used to reduce the oscillation of the laser scanner at resonance frequency. To design a qualified controller, the model of the laser scanner is set up. The transfer function of the model is identified with MATLAB according to the tested data. After introducing of the controller, the nonlinearity decreases from 13.75% to 2.67% at 50 Hz. The laser scanner also has other advantages such as large deflection mirror, small mechanical structure, and high scanning speed.
Rothacker, Karen M; Brown, Suzanne J; Hadlow, Narelle C; Wardrop, Robert; Walsh, John P
2016-03-01
The TSH-T4 relationship was thought to be inverse log-linear, but recent cross-sectional studies report a complex, nonlinear relationship; large, intra-individual studies are lacking. Our objective was to analyze the TSH-free T4 relationship within individuals. We analyzed data from 13 379 patients, each with six or more TSH/free T4 measurements and at least a 5-fold difference between individual median TSH and minimum or maximum TSH. Linear and nonlinear regression models of log TSH on free T4 were fitted to data from individuals and goodness of fit compared by likelihood ratio testing. Comparing all models, the linear model achieved best fit in 31% of individuals, followed by quartic (27%), cubic (15%), null (12%), and quadratic (11%) models. After eliminating least favored models (with individuals reassigned to best fitting, available models), the linear model fit best in 42% of participants, quartic in 43%, and null model in 15%. As the number of observations per individual increased, so did the proportion of individuals in whom the linear model achieved best fit, to 66% in those with more than 20 observations. When linear models were applied to all individuals and averaged according to individual median free T4 values, variations in slope and intercept indicated a nonlinear log TSH-free T4 relationship across the population. The log TSH-free T4 relationship appears linear in some individuals and nonlinear in others, but is predominantly linear in those with the largest number of observations. A log-linear relationship within individuals can be reconciled with a non-log-linear relationship in a population.
International Nuclear Information System (INIS)
Wichmann, K.
2009-01-01
Recently, Letters of Intent (LoI) for experiments at the International Linear Collider (ILC) have been submitted. Among the three proposals is the International Large Detector (ILD) concept which is at the focus of these studies. From various subjects addressed in the LoI, a wide spectrum of studies of SUSY particle properties is presented here. Most of them are benchmark reactions for the ILC and can be used both in physics studies and in work on detector design and optimization, respectively. All studies were performed with a full detector simulation using GEANT4, which is a great improvement compared to the previous results with much less detailed, so called f ast , simulation (SIMDET). The importance of this improved simulation is reflected in the results. The presented analyzes have been chosen to be the most challenging for the detector to study its performance and guide the detector development. Additionally an important problem of unavoidable beam induced backgrounds at linear colliders is addressed and ways of reducing its impact on physics studies are shown for an example SUSY analysis. (author)
Low-impedance internal linear inductive antenna for large-area flat panel display plasma processing
International Nuclear Information System (INIS)
Kim, K.N.; Jung, S.J.; Lee, Y.J.; Yeom, G.Y.; Lee, S.H.; Lee, J.K.
2005-01-01
An internal-type linear inductive antenna, that is, a double-comb-type antenna, was developed for a large-area plasma source having the size of 1020 mmx830 mm, and high density plasmas on the order of 2.3x10 11 cm -3 were obtained with 15 mTorr Ar at 5000 W of inductive power with good plasma stability. This is higher than that for the conventional serpentine-type antenna, possibly due to the low impedance, resulting in high efficiency of power transfer for the double-comb antenna type. In addition, due to the remarkable reduction of the antenna length, a plasma uniformity of less than 8% was obtained within the substrate area of 880 mmx660 mm at 5000 W without having a standing-wave effect
Linear velocity fields in non-Gaussian models for large-scale structure
Scherrer, Robert J.
1992-01-01
Linear velocity fields in two types of physically motivated non-Gaussian models are examined for large-scale structure: seed models, in which the density field is a convolution of a density profile with a distribution of points, and local non-Gaussian fields, derived from a local nonlinear transformation on a Gaussian field. The distribution of a single component of the velocity is derived for seed models with randomly distributed seeds, and these results are applied to the seeded hot dark matter model and the global texture model with cold dark matter. An expression for the distribution of a single component of the velocity in arbitrary local non-Gaussian models is given, and these results are applied to such fields with chi-squared and lognormal distributions. It is shown that all seed models with randomly distributed seeds and all local non-Guassian models have single-component velocity distributions with positive kurtosis.
Non-linear vibrating systems excited by a nonideal energy source with a large slope characteristic
González-Carbajal, Javier; Domínguez, Jaime
2017-11-01
This paper revisits the problem of an unbalanced motor attached to a fixed frame by means of a nonlinear spring and a linear damper. The excitation provided by the motor is, in general, nonideal, which means it is affected by the vibratory response. Since the system behaviour is highly dependent on the order of magnitude of the motor characteristic slope, the case of large slope is considered herein. Some Perturbation Methods are applied to the system of equations, which allows transforming the original 4D system into a much simpler 2D system. The fixed points of this reduced system and their stability are carefully studied. We find the existence of a Hopf bifurcation which, to the authors' knowledge, has not been addressed before in the literature. These analytical results are supported by numerical simulations. We also compare our approach and results with those published by other authors.
Flexible non-linear predictive models for large-scale wind turbine diagnostics
DEFF Research Database (Denmark)
Bach-Andersen, Martin; Rømer-Odgaard, Bo; Winther, Ole
2017-01-01
We demonstrate how flexible non-linear models can provide accurate and robust predictions on turbine component temperature sensor data using data-driven principles and only a minimum of system modeling. The merits of different model architectures are evaluated using data from a large set...... of turbines operating under diverse conditions. We then go on to test the predictive models in a diagnostic setting, where the output of the models are used to detect mechanical faults in rotor bearings. Using retrospective data from 22 actual rotor bearing failures, the fault detection performance...... of the models are quantified using a structured framework that provides the metrics required for evaluating the performance in a fleet wide monitoring setup. It is demonstrated that faults are identified with high accuracy up to 45 days before a warning from the hard-threshold warning system....
Cosmological large-scale structures beyond linear theory in modified gravity
Energy Technology Data Exchange (ETDEWEB)
Bernardeau, Francis; Brax, Philippe, E-mail: francis.bernardeau@cea.fr, E-mail: philippe.brax@cea.fr [CEA, Institut de Physique Théorique, 91191 Gif-sur-Yvette Cédex (France)
2011-06-01
We consider the effect of modified gravity on the growth of large-scale structures at second order in perturbation theory. We show that modified gravity models changing the linear growth rate of fluctuations are also bound to change, although mildly, the mode coupling amplitude in the density and reduced velocity fields. We present explicit formulae which describe this effect. We then focus on models of modified gravity involving a scalar field coupled to matter, in particular chameleons and dilatons, where it is shown that there exists a transition scale around which the existence of an extra scalar degree of freedom induces significant changes in the coupling properties of the cosmic fields. We obtain the amplitude of this effect for realistic dilaton models at the tree-order level for the bispectrum, finding them to be comparable in amplitude to those obtained in the DGP and f(R) models.
Enhanced 2D-DOA Estimation for Large Spacing Three-Parallel Uniform Linear Arrays
Directory of Open Access Journals (Sweden)
Dong Zhang
2018-01-01
Full Text Available An enhanced two-dimensional direction of arrival (2D-DOA estimation algorithm for large spacing three-parallel uniform linear arrays (ULAs is proposed in this paper. Firstly, we use the propagator method (PM to get the highly accurate but ambiguous estimation of directional cosine. Then, we use the relationship between the directional cosine to eliminate the ambiguity. This algorithm not only can make use of the elements of the three-parallel ULAs but also can utilize the connection between directional cosine to improve the estimation accuracy. Besides, it has satisfied estimation performance when the elevation angle is between 70° and 90° and it can automatically pair the estimated azimuth and elevation angles. Furthermore, it has low complexity without using any eigen value decomposition (EVD or singular value decompostion (SVD to the covariance matrix. Simulation results demonstrate the effectiveness of our proposed algorithm.
Directory of Open Access Journals (Sweden)
Ichitaro Yamazaki
2015-01-01
of their low-rank properties. To compute a low-rank approximation of a dense matrix, in this paper, we study the performance of QR factorization with column pivoting or with restricted pivoting on multicore CPUs with a GPU. We first propose several techniques to reduce the postprocessing time, which is required for restricted pivoting, on a modern CPU. We then examine the potential of using a GPU to accelerate the factorization process with both column and restricted pivoting. Our performance results on two eight-core Intel Sandy Bridge CPUs with one NVIDIA Kepler GPU demonstrate that using the GPU, the factorization time can be reduced by a factor of more than two. In addition, to study the performance of our implementations in practice, we integrate them into a recently developed software StruMF which algebraically exploits such low-rank structures for solving a general sparse linear system of equations. Our performance results for solving Poisson's equations demonstrate that the proposed techniques can significantly reduce the preconditioner construction time of StruMF on the CPUs, and the construction time can be further reduced by 10%–50% using the GPU.
Ferencz, Donald C.; Viterna, Larry A.
1991-01-01
ALPS is a computer program which can be used to solve general linear program (optimization) problems. ALPS was designed for those who have minimal linear programming (LP) knowledge and features a menu-driven scheme to guide the user through the process of creating and solving LP formulations. Once created, the problems can be edited and stored in standard DOS ASCII files to provide portability to various word processors or even other linear programming packages. Unlike many math-oriented LP solvers, ALPS contains an LP parser that reads through the LP formulation and reports several types of errors to the user. ALPS provides a large amount of solution data which is often useful in problem solving. In addition to pure linear programs, ALPS can solve for integer, mixed integer, and binary type problems. Pure linear programs are solved with the revised simplex method. Integer or mixed integer programs are solved initially with the revised simplex, and the completed using the branch-and-bound technique. Binary programs are solved with the method of implicit enumeration. This manual describes how to use ALPS to create, edit, and solve linear programming problems. Instructions for installing ALPS on a PC compatible computer are included in the appendices along with a general introduction to linear programming. A programmers guide is also included for assistance in modifying and maintaining the program.
Wang, Wenhong
2013-07-12
We report the observation of a large linear magnetoresistance (MR) and Shubnikov-de Hass (SdH) quantum oscillations in single crystals of YPdBi Heusler topological insulators. Owning to the successfully obtained the high-quality YPdBi single crystals, large non-saturating linear MR of as high as 350% at 5K and over 120% at 300K under a moderate magnetic field of 7T is observed. In addition to the large, field-linear MR, the samples exhibit pronounced SdH quantum oscillations at low temperature. Analysis of the SdH data manifests that the high-mobility bulk electron carriers dominate the magnetotransport and are responsible for the observed large linear MR in YPdBi crystals. These findings imply that the Heusler-based topological insulators have superiorities for investigating the novel quantum transport properties and developing the potential applications.
Wang, Wenhong; Du, Yin; Xu, Guizhou; Zhang, Xiaoming; Liu, Enke; Liu, Zhongyuan; Shi, Youguo; Chen, Jinglan; Wu, Guangheng; Zhang, Xixiang
2013-01-01
We report the observation of a large linear magnetoresistance (MR) and Shubnikov-de Hass (SdH) quantum oscillations in single crystals of YPdBi Heusler topological insulators. Owning to the successfully obtained the high-quality YPdBi single crystals, large non-saturating linear MR of as high as 350% at 5K and over 120% at 300K under a moderate magnetic field of 7T is observed. In addition to the large, field-linear MR, the samples exhibit pronounced SdH quantum oscillations at low temperature. Analysis of the SdH data manifests that the high-mobility bulk electron carriers dominate the magnetotransport and are responsible for the observed large linear MR in YPdBi crystals. These findings imply that the Heusler-based topological insulators have superiorities for investigating the novel quantum transport properties and developing the potential applications.
Energy Technology Data Exchange (ETDEWEB)
Carey, G.F.; Young, D.M.
1993-12-31
The program outlined here is directed to research on methods, algorithms, and software for distributed parallel supercomputers. Of particular interest are finite element methods and finite difference methods together with sparse iterative solution schemes for scientific and engineering computations of very large-scale systems. Both linear and nonlinear problems will be investigated. In the nonlinear case, applications with bifurcation to multiple solutions will be considered using continuation strategies. The parallelizable numerical methods of particular interest are a family of partitioning schemes embracing domain decomposition, element-by-element strategies, and multi-level techniques. The methods will be further developed incorporating parallel iterative solution algorithms with associated preconditioners in parallel computer software. The schemes will be implemented on distributed memory parallel architectures such as the CRAY MPP, Intel Paragon, the NCUBE3, and the Connection Machine. We will also consider other new architectures such as the Kendall-Square (KSQ) and proposed machines such as the TERA. The applications will focus on large-scale three-dimensional nonlinear flow and reservoir problems with strong convective transport contributions. These are legitimate grand challenge class computational fluid dynamics (CFD) problems of significant practical interest to DOE. The methods developed and algorithms will, however, be of wider interest.
Fast Kalman-like filtering for large-dimensional linear and Gaussian state-space models
Ait-El-Fquih, Boujemaa; Hoteit, Ibrahim
2015-01-01
This paper considers the filtering problem for linear and Gaussian state-space models with large dimensions, a setup in which the optimal Kalman Filter (KF) might not be applicable owing to the excessive cost of manipulating huge covariance matrices. Among the most popular alternatives that enable cheaper and reasonable computation is the Ensemble KF (EnKF), a Monte Carlo-based approximation. In this paper, we consider a class of a posteriori distributions with diagonal covariance matrices and propose fast approximate deterministic-based algorithms based on the Variational Bayesian (VB) approach. More specifically, we derive two iterative KF-like algorithms that differ in the way they operate between two successive filtering estimates; one involves a smoothing estimate and the other involves a prediction estimate. Despite its iterative nature, the prediction-based algorithm provides a computational cost that is, on the one hand, independent of the number of iterations in the limit of very large state dimensions, and on the other hand, always much smaller than the cost of the EnKF. The cost of the smoothing-based algorithm depends on the number of iterations that may, in some situations, make this algorithm slower than the EnKF. The performances of the proposed filters are studied and compared to those of the KF and EnKF through a numerical example.
Directory of Open Access Journals (Sweden)
Chun-Fu Chen
2014-03-01
Full Text Available Linear analytical study on the mechanical sensitivity in large deflection of unsymmetrically layered and laterally loaded piezoelectric plate under pretension is conducted. von Karman plate theory for large deflection is utilized but extended to the case of an unsymmetrically layered plate embedded with a piezoelectric layer. The governing equations thus obtained are simplified by omitting the arising nonlinear terms, yielding a Bessel or modified Bessel equation for the lateral slope. Depending on the relative magnitude of the piezoelectric effect, for both cases, analytical solutions of various geometrical responses are developed and formulated via Bessel and modified Bessel functions. The associated ultimate radial stresses are further derived following lamina constitutive law to evaluate the mechanical sensitivity of the considered plate. For a nearly monolithic plate under a very low applied voltage, the results are in good agreement with those for a single-layered case due to pure mechanical load available in literature, and thus the present approach is checked. For a two-layered unsymmetric plate made of typical silicon-based materials, a sound piezoelectric effect is illustrated particularly in a low pretension condition.
Narimani, Zahra; Beigy, Hamid; Ahmad, Ashar; Masoudi-Nejad, Ali; Fröhlich, Holger
2017-01-01
Inferring the structure of molecular networks from time series protein or gene expression data provides valuable information about the complex biological processes of the cell. Causal network structure inference has been approached using different methods in the past. Most causal network inference techniques, such as Dynamic Bayesian Networks and ordinary differential equations, are limited by their computational complexity and thus make large scale inference infeasible. This is specifically true if a Bayesian framework is applied in order to deal with the unavoidable uncertainty about the correct model. We devise a novel Bayesian network reverse engineering approach using ordinary differential equations with the ability to include non-linearity. Besides modeling arbitrary, possibly combinatorial and time dependent perturbations with unknown targets, one of our main contributions is the use of Expectation Propagation, an algorithm for approximate Bayesian inference over large scale network structures in short computation time. We further explore the possibility of integrating prior knowledge into network inference. We evaluate the proposed model on DREAM4 and DREAM8 data and find it competitive against several state-of-the-art existing network inference methods.
Fast Kalman-like filtering for large-dimensional linear and Gaussian state-space models
Ait-El-Fquih, Boujemaa
2015-08-13
This paper considers the filtering problem for linear and Gaussian state-space models with large dimensions, a setup in which the optimal Kalman Filter (KF) might not be applicable owing to the excessive cost of manipulating huge covariance matrices. Among the most popular alternatives that enable cheaper and reasonable computation is the Ensemble KF (EnKF), a Monte Carlo-based approximation. In this paper, we consider a class of a posteriori distributions with diagonal covariance matrices and propose fast approximate deterministic-based algorithms based on the Variational Bayesian (VB) approach. More specifically, we derive two iterative KF-like algorithms that differ in the way they operate between two successive filtering estimates; one involves a smoothing estimate and the other involves a prediction estimate. Despite its iterative nature, the prediction-based algorithm provides a computational cost that is, on the one hand, independent of the number of iterations in the limit of very large state dimensions, and on the other hand, always much smaller than the cost of the EnKF. The cost of the smoothing-based algorithm depends on the number of iterations that may, in some situations, make this algorithm slower than the EnKF. The performances of the proposed filters are studied and compared to those of the KF and EnKF through a numerical example.
Extension of the linear nodal method to large concrete building calculations
International Nuclear Information System (INIS)
Childs, R.L.; Rhoades, W.A.
1985-01-01
The implementation of the linear nodal method in the TORT code is described, and the results of a mesh refinement study to test the effectiveness of the linear nodal and weighted diamond difference methods available in TORT are presented
Prototyping a large field size IORT applicator for a mobile linear accelerator
Energy Technology Data Exchange (ETDEWEB)
Janssen, Rogier W J; Dries, Wim J F [Catharina-Hospital Eindhoven, PO Box 1350, 5602 ZA, Eindhoven (Netherlands); Faddegon, Bruce A [University of California San Francisco Comprehensive Cancer Center, 1600 Divisadero Street, San Francisco, CA 94115-1708 (United States)], E-mail: rogier.janssen@mac.com
2008-04-21
The treatment of large tumors such as sarcomas with intra-operative radiotherapy using a Mobetron (registered) is often complicated because of the limited field size of the primary collimator and the available applicators (max Oe100 mm). To circumvent this limitation a prototype rectangular applicator of 80 x 150 mm{sup 2} was designed and built featuring an additional scattering foil located at the top of the applicator. Because of its proven accuracy in modeling linear accelerator components the design was based on the EGSnrc Monte Carlo simulation code BEAMnrc. First, the Mobetron (registered) treatment head was simulated both without an applicator and with a standard 100 mm applicator. Next, this model was used to design an applicator foil consisting of a rectangular Al base plate covering the whole beam and a pyramid of four stacked cylindrical slabs of different diameters centered on top of it. This foil was mounted on top of a plain rectangular Al tube. A prototype was built and tested with diode dosimetry in a water tank. Here, the prototype showed clinically acceptable 80 x 150 mm{sup 2} dose distributions for 4 MeV, 6 MeV and 9 MeV, obviating the use of complicated multiple irradiations with abutting field techniques. In addition, the measurements agreed well with the MC simulations, typically within 2%/1 mm.
Prototyping a large field size IORT applicator for a mobile linear accelerator
International Nuclear Information System (INIS)
Janssen, Rogier W J; Dries, Wim J F; Faddegon, Bruce A
2008-01-01
The treatment of large tumors such as sarcomas with intra-operative radiotherapy using a Mobetron (registered) is often complicated because of the limited field size of the primary collimator and the available applicators (max Oe100 mm). To circumvent this limitation a prototype rectangular applicator of 80 x 150 mm 2 was designed and built featuring an additional scattering foil located at the top of the applicator. Because of its proven accuracy in modeling linear accelerator components the design was based on the EGSnrc Monte Carlo simulation code BEAMnrc. First, the Mobetron (registered) treatment head was simulated both without an applicator and with a standard 100 mm applicator. Next, this model was used to design an applicator foil consisting of a rectangular Al base plate covering the whole beam and a pyramid of four stacked cylindrical slabs of different diameters centered on top of it. This foil was mounted on top of a plain rectangular Al tube. A prototype was built and tested with diode dosimetry in a water tank. Here, the prototype showed clinically acceptable 80 x 150 mm 2 dose distributions for 4 MeV, 6 MeV and 9 MeV, obviating the use of complicated multiple irradiations with abutting field techniques. In addition, the measurements agreed well with the MC simulations, typically within 2%/1 mm
Large linear magnetoresistance in a new Dirac material BaMnBi2
Wang, Yi-Yan; Yu, Qiao-He; Xia, Tian-Long
2016-10-01
Dirac semimetal is a class of materials that host Dirac fermions as emergent quasi-particles. Dirac cone-type band structure can bring interesting properties such as quantum linear magnetoresistance and large mobility in the materials. In this paper, we report the synthesis of high quality single crystals of BaMnBi2 and investigate the transport properties of the samples. BaMnBi2 is a metal with an antiferromagnetic transition at T N = 288 K. The temperature dependence of magnetization displays different behavior from CaMnBi2 and SrMnBi2, which suggests the possible different magnetic structure of BaMnBi2. The Hall data reveals electron-type carriers and a mobility μ(5 K) = 1500 cm2/V·s. Angle-dependent magnetoresistance reveals the quasi-two-dimensional (2D) Fermi surface in BaMnBi2. A crossover from semiclassical MR ˜ H 2 dependence in low field to MR ˜ H dependence in high field, which is attributed to the quantum limit of Dirac fermions, has been observed in magnetoresistance. Our results indicate the existence of Dirac fermions in BaMnBi2. Project supported by the National Natural Science Foundation of China (Grant No. 11574391), the Fundamental Research Funds for the Central Universities, and the Research Funds of Renmin University of China (Grant No. 14XNLQ07).
Large linear magnetoresistance in a new Dirac material BaMnBi2
International Nuclear Information System (INIS)
Wang Yi-Yan; Yu Qiao-He; Xia Tian-Long
2016-01-01
Dirac semimetal is a class of materials that host Dirac fermions as emergent quasi-particles. Dirac cone-type band structure can bring interesting properties such as quantum linear magnetoresistance and large mobility in the materials. In this paper, we report the synthesis of high quality single crystals of BaMnBi 2 and investigate the transport properties of the samples. BaMnBi 2 is a metal with an antiferromagnetic transition at T N = 288 K. The temperature dependence of magnetization displays different behavior from CaMnBi 2 and SrMnBi 2 , which suggests the possible different magnetic structure of BaMnBi 2 . The Hall data reveals electron-type carriers and a mobility μ (5 K) = 1500 cm 2 /V·s. Angle-dependent magnetoresistance reveals the quasi-two-dimensional (2D) Fermi surface in BaMnBi 2 . A crossover from semiclassical MR ∼ H 2 dependence in low field to MR ∼ H dependence in high field, which is attributed to the quantum limit of Dirac fermions, has been observed in magnetoresistance. Our results indicate the existence of Dirac fermions in BaMnBi 2 . (rapid communication)
Triple Z0-Boson Production in a Large Extra Dimensions Model at the International Linear Collider
International Nuclear Information System (INIS)
Jiang Ruo-Cheng; Li Xiao-Zhou; Ma Wen-Gan; Guo Lei; Zhang Ren-You
2012-01-01
We investigate the effects induced by the interactions of the Kaluza—Klein graviton with the standard model (SM) particles on the triple Z 0 -boson production process at the International Linear Collider in the framework of the large extra dimension (LED) model. We present the dependence of the integrated cross sections on the electron-positron colliding energy √s, and various kinematic distributions of final Z 0 bosons and their subsequential decay products in both the SM and the LED model. We also provide the relationship between the integrated cross section and the fundamental scale MS by taking the number of the extra dimensions (d) as 3, 4, 5, and 6, respectively. The numerical results show that the LED effect can induce an observable relative discrepancy for the integrated cross section (δ LED ). We find that the relative discrepancy of the LED effect can even reach a few dozen percent in the high transverse momentum area or the central rapidity region of the final Z 0 -bosons and muons
The linearly scaling 3D fragment method for large scale electronic structure calculations
Energy Technology Data Exchange (ETDEWEB)
Zhao Zhengji [National Energy Research Scientific Computing Center (NERSC) (United States); Meza, Juan; Shan Hongzhang; Strohmaier, Erich; Bailey, David; Wang Linwang [Computational Research Division, Lawrence Berkeley National Laboratory (United States); Lee, Byounghak, E-mail: ZZhao@lbl.go [Physics Department, Texas State University (United States)
2009-07-01
The linearly scaling three-dimensional fragment (LS3DF) method is an O(N) ab initio electronic structure method for large-scale nano material simulations. It is a divide-and-conquer approach with a novel patching scheme that effectively cancels out the artificial boundary effects, which exist in all divide-and-conquer schemes. This method has made ab initio simulations of thousand-atom nanosystems feasible in a couple of hours, while retaining essentially the same accuracy as the direct calculation methods. The LS3DF method won the 2008 ACM Gordon Bell Prize for algorithm innovation. Our code has reached 442 Tflop/s running on 147,456 processors on the Cray XT5 (Jaguar) at OLCF, and has been run on 163,840 processors on the Blue Gene/P (Intrepid) at ALCF, and has been applied to a system containing 36,000 atoms. In this paper, we will present the recent parallel performance results of this code, and will apply the method to asymmetric CdSe/CdS core/shell nanorods, which have potential applications in electronic devices and solar cells.
Vasilev, Aleksandr S.; Konyakhin, Igor A.; Timofeev, Alexander N.; Lashmanov, Oleg U.; Molev, Fedor V.
2015-05-01
The paper analyzes the construction matters and metrological parameters of the electrooptic converter to control linear displacements of the large structures of the buildings and facilities. The converter includes the base module, the processing module and a set of the reference marks. The base module is the main unit of the system, it includes the receiving optical system and the CMOS photodetector array that realizes the instrument coordinate system that controls the mark coordinates in the space. The methods of the frame-to-frame difference, adaptive threshold filtration, binarization and objects search by the tied areas to detect the marks against accidental contrast background is the basis of the algorithm. The entire algorithm is performed during one image reading stage and is based on the FPGA. The developed and manufactured converter experimental model was tested in laboratory conditions at the metrological bench at the distance between the base module and the mark 50±0.2 m. The static characteristic was read during the experiment of the reference mark displacement at the pitch of 5 mm in the horizontal and vertical directions for the displacement range 400 mm. The converter experimental model error not exceeding ±0.5 mm was obtained in the result of the experiment.
Low temperature diamond growth by linear antenna plasma CVD over large area
International Nuclear Information System (INIS)
Izak, Tibor; Babchenko, Oleg; Potocky, Stepan; Kromka, Alexander; Varga, Marian
2012-01-01
Recently, there is a great effort to increase the deposition area and decrease the process temperature for diamond growth which will enlarge its applications including use of temperature sensitive substrates. In this work, we report on the large area (20 x 30 cm 2 ) and low temperature (250 C) polycrystalline diamond growth by pulsed linear antenna microwave plasma system. The influence of substrate temperature varied from 250 to 680 C, as controlled by the table heater and/or by microwave power, is studied. It was found that the growth rate, film morphology and diamond to non-diamond phases (sp 3 /sp 2 carbon bonds) are influenced by the growth temperature, as confirmed by SEM and Raman measurements. The surface chemistry and growth processes were studied in terms of activation energies (E a ) calculated from Arrhenius plots. The activation energies of growth processes were very low (1.7 and 7.8 kcal mol -1 ) indicating an energetically favourable growth process from the CO 2 -CH 4 -H 2 gas mixture. In addition, from activation energies two different growth regimes were observed at low and high temperatures, indicating different growth mechanism. (Copyright copyright 2012 WILEY-VCH Verlag GmbH and Co. KGaA, Weinheim)
Energy Technology Data Exchange (ETDEWEB)
Clemens, M.; Weiland, T. [Technische Hochschule Darmstadt (Germany)
1996-12-31
In the field of computational electrodynamics the discretization of Maxwell`s equations using the Finite Integration Theory (FIT) yields very large, sparse, complex symmetric linear systems of equations. For this class of complex non-Hermitian systems a number of conjugate gradient-type algorithms is considered. The complex version of the biconjugate gradient (BiCG) method by Jacobs can be extended to a whole class of methods for complex-symmetric algorithms SCBiCG(T, n), which only require one matrix vector multiplication per iteration step. In this class the well-known conjugate orthogonal conjugate gradient (COCG) method for complex-symmetric systems corresponds to the case n = 0. The case n = 1 yields the BiCGCR method which corresponds to the conjugate residual algorithm for the real-valued case. These methods in combination with a minimal residual smoothing process are applied separately to practical 3D electro-quasistatical and eddy-current problems in electrodynamics. The practical performance of the SCBiCG methods is compared with other methods such as QMR and TFQMR.
On Feature Extraction from Large Scale Linear LiDAR Data
Acharjee, Partha Pratim
Airborne light detection and ranging (LiDAR) can generate co-registered elevation and intensity map over large terrain. The co-registered 3D map and intensity information can be used efficiently for different feature extraction application. In this dissertation, we developed two algorithms for feature extraction, and usages of features for practical applications. One of the developed algorithms can map still and flowing waterbody features, and another one can extract building feature and estimate solar potential on rooftops and facades. Remote sensing capabilities, distinguishing characteristics of laser returns from water surface and specific data collection procedures provide LiDAR data an edge in this application domain. Furthermore, water surface mapping solutions must work on extremely large datasets, from a thousand square miles, to hundreds of thousands of square miles. National and state-wide map generation/upgradation and hydro-flattening of LiDAR data for many other applications are two leading needs of water surface mapping. These call for as much automation as possible. Researchers have developed many semi-automated algorithms using multiple semi-automated tools and human interventions. This reported work describes a consolidated algorithm and toolbox developed for large scale, automated water surface mapping. Geometric features such as flatness of water surface, higher elevation change in water-land interface and, optical properties such as dropouts caused by specular reflection, bimodal intensity distributions were some of the linear LiDAR features exploited for water surface mapping. Large-scale data handling capabilities are incorporated by automated and intelligent windowing, by resolving boundary issues and integrating all results to a single output. This whole algorithm is developed as an ArcGIS toolbox using Python libraries. Testing and validation are performed on a large datasets to determine the effectiveness of the toolbox and results are
Joint shape segmentation with linear programming
Huang, Qixing; Koltun, Vladlen; Guibas, Leonidas
2011-01-01
program is solved via a linear programming relaxation, using a block coordinate descent procedure that makes the optimization feasible for large databases. We evaluate the presented approach on the Princeton segmentation benchmark and show that joint shape
Mashood, K. K.; Singh, Vijay A.
2013-01-01
Research suggests that problem-solving skills are transferable across domains. This claim, however, needs further empirical substantiation. We suggest correlation studies as a methodology for making preliminary inferences about transfer. The correlation of the physics performance of students with their performance in chemistry and mathematics in…
Ponomarev, A. L.; Brenner, D.; Hlatky, L. R.; Sachs, R. K.
2000-01-01
DNA double-strand breaks (DSBs) produced by densely ionizing radiation are not located randomly in the genome: recent data indicate DSB clustering along chromosomes. Stochastic DSB clustering at large scales, from > 100 Mbp down to simulations and analytic equations. A random-walk, coarse-grained polymer model for chromatin is combined with a simple track structure model in Monte Carlo software called DNAbreak and is applied to data on alpha-particle irradiation of V-79 cells. The chromatin model neglects molecular details but systematically incorporates an increase in average spatial separation between two DNA loci as the number of base-pairs between the loci increases. Fragment-size distributions obtained using DNAbreak match data on large fragments about as well as distributions previously obtained with a less mechanistic approach. Dose-response relations, linear at small doses of high linear energy transfer (LET) radiation, are obtained. They are found to be non-linear when the dose becomes so large that there is a significant probability of overlapping or close juxtaposition, along one chromosome, for different DSB clusters from different tracks. The non-linearity is more evident for large fragments than for small. The DNAbreak results furnish an example of the RLC (randomly located clusters) analytic formalism, which generalizes the broken-stick fragment-size distribution of the random-breakage model that is often applied to low-LET data.
DEFF Research Database (Denmark)
Knudsen, Jesper Viese; Bendtsen, Jan Dimon; Andersen, Palle
2016-01-01
In this paper, a self-tuning linear quadratic supervisory regulator using a large-signal state estimator for a diesel driven generator set is proposed. The regulator improves operational efficiency, in comparison to current implementations, by (i) automating the initial tuning process and (ii...... throughout the operating range of the diesel generator....
Camus, Marine; Jensen, Dennis M.; Ohning, Gordon V.; Kovacs, Thomas O.; Ghassemi, Kevin A.; Jutabha, Rome; Machicado, Gustavo A.; Dulai, Gareth S.; Hines, Joel O.
2013-01-01
Background and study aims Cameron ulcers are a rare but clinically significant cause of severe upper gastrointestinal hemorrhage (SUGIH). Our aims were to describe (1) the diagnosis, treatment and outcomes of patients with Cameron ulcers causing hospitalization for SUGIH, (2) the differences between patients with occult vs. overt bleeding and (3) between patients treated surgically and medically. Patients and methods Over the past 17 years, all consecutive patients hospitalized in our two tertiary referral medical centers for severe UGIH or severe obscure GIH and entered into our large prospective databasis were screened for Cameron ulcer diagnosis. Results Cameron ulcers were diagnosed in 25 patients of 3960 patients with SUGIH (0.6%). 21 patients had follow-up (median [IQR] time of 20.4 months [8.5–31.8]). Patients were more often elderly females with chronic anemia, always had large hiatal hernias, and were usually referred for obscure SUGIH. Twelve (57.2%) patients were referred to surgery for rebleeding and recurrent blood loss while treated with high dose of proton pump inhibitors (PPI). 9 (42.8%) other patients continued PPI without any rebleeding during the follow-up. Patients with overt bleeding had significantly more prior hospitalizations for SUGIH, more often stigmata of hemorrhage on ulcers, and more red blood cell transfusions than patients with occult bleeding. However, there was no difference in rebleeding and mortality rates between the two groups. Conclusions Cameron ulcers in large hiatal hernias are an uncommon cause of SUGIH. Most of patients are referred for obscure GIH. The choice of medical vs. surgical therapy should be individualized. PMID:23616128
Simplified Linear Equation Solvers users manual
Energy Technology Data Exchange (ETDEWEB)
Gropp, W. [Argonne National Lab., IL (United States); Smith, B. [California Univ., Los Angeles, CA (United States)
1993-02-01
The solution of large sparse systems of linear equations is at the heart of many algorithms in scientific computing. The SLES package is a set of easy-to-use yet powerful and extensible routines for solving large sparse linear systems. The design of the package allows new techniques to be used in existing applications without any source code changes in the applications.
A Linearized Large Signal Model of an LCL-Type Resonant Converter
Directory of Open Access Journals (Sweden)
Hong-Yu Li
2015-03-01
Full Text Available In this work, an LCL-type resonant dc/dc converter with a capacitive output filter is modeled in two stages. In the first high-frequency ac stage, all ac signals are decomposed into two orthogonal vectors in a synchronous rotating d–q frame using multi-frequency modeling. In the dc stage, all dc quantities are represented by their average values with average state-space modeling. A nonlinear two-stage model is then created by means of a non-linear link. By aligning the transformer voltage on the d-axis, the nonlinear link can be eliminated, and the whole converter can be modeled by a single set of linear state-space equations. Furthermore, a feedback control scheme can be formed according to the steady-state solutions. Simulation and experimental results have proven that the resulted model is good for fast simulation and state variable estimation.
McCaskill, John
There can be large spatial and temporal separation of cause and effect in policy making. Determining the correct linkage between policy inputs and outcomes can be highly impractical in the complex environments faced by policy makers. In attempting to see and plan for the probable outcomes, standard linear models often overlook, ignore, or are unable to predict catastrophic events that only seem improbable due to the issue of multiple feedback loops. There are several issues with the makeup and behaviors of complex systems that explain the difficulty many mathematical models (factor analysis/structural equation modeling) have in dealing with non-linear effects in complex systems. This chapter highlights those problem issues and offers insights to the usefulness of ABM in dealing with non-linear effects in complex policy making environments.
Klegeris, Andis; Bahniwal, Manpreet; Hurren, Heather
2013-01-01
Problem-based learning (PBL) was originally introduced in medical education programs as a form of small-group learning, but its use has now spread to large undergraduate classrooms in various other disciplines. Introduction of new teaching techniques, including PBL-based methods, needs to be justified by demonstrating the benefits of such techniques over classical teaching styles. Previously, we demonstrated that introduction of tutor-less PBL in a large third-year biochemistry undergraduate class increased student satisfaction and attendance. The current study assessed the generic problem-solving abilities of students from the same class at the beginning and end of the term, and compared student scores with similar data obtained in three classes not using PBL. Two generic problem-solving tests of equal difficulty were administered such that students took different tests at the beginning and the end of the term. Blinded marking showed a statistically significant 13% increase in the test scores of the biochemistry students exposed to PBL, while no trend toward significant change in scores was observed in any of the control groups not using PBL. Our study is among the first to demonstrate that use of tutor-less PBL in a large classroom leads to statistically significant improvement in generic problem-solving skills of students. PMID:23463230
International Nuclear Information System (INIS)
Lee, Hwang; Kok, Pieter; Dowling, Jonathan P.; Cerf, Nicolas J.
2002-01-01
We propose a method for preparing maximal path entanglement with a definite photon-number N, larger than two, using projective measurements. In contrast with the previously known schemes, our method uses only linear optics. Specifically, we exhibit a way of generating four-photon, path-entangled states of the form vertical bar 4,0>+ vertical bar 0,4>, using only four beam splitters and two detectors. These states are of major interest as a resource for quantum interferometric sensors as well as for optical quantum lithography and quantum holography
A large-scale linear complementarity model of the North American natural gas market
International Nuclear Information System (INIS)
Gabriel, Steven A.; Jifang Zhuang; Kiet, Supat
2005-01-01
The North American natural gas market has seen significant changes recently due to deregulation and restructuring. For example, third party marketers can contract for transportation and purchase of gas to sell to end-users. While the intent was a more competitive market, the potential for market power exists. We analyze this market using a linear complementarity equilibrium model including producers, storage and peak gas operators, third party marketers and four end-use sectors. The marketers are depicted as Nash-Cournot players determining supply to meet end-use consumption, all other players are in perfect competition. Results based on National Petroleum Council scenarios are presented. (Author)
Definition of a reference metrology network for the positioning of a large linear accelerator
International Nuclear Information System (INIS)
Becker, F.
2003-12-01
This thesis is a study of the Compact Linear Collider (CLIC) alignment system, a project of linear accelerator of about 30 km long of the European Organization for Nuclear Research (CERN). The pre-alignment tolerance on the transverse positions of the components of the CLIC linacs is typically ten microns over distances of 200 m. This research is a consequence of 10 years work, where several sets of special sensors dedicated to metrology have been adapted for the CLIC project. Most of these sensors deliver measurements linked to geometric references sensitive to gravity fluctuation. An important part of this work is therefore dedicated to study the gravity disruptions as a high level of accuracy is required. The parameters to take into account in the use of the hydrostatic leveling have thus been highlighted. A proposal of configuration of the system alignment based on a selection of sensors has also been given in this research. Computer models of different possible configurations have been presented. As the existing computing software was inappropriate, a new object oriented software package has been developed, to ensure future upgrades. An optimized configuration of the network has been defined from a set of simulations. Finally, due to problems in the use of hydrostatic leveling systems, a solution based on the use of a long laser beam as an alternative solution is discussed. (author)
Imprint of non-linear effects on HI intensity mapping on large scales
Energy Technology Data Exchange (ETDEWEB)
Umeh, Obinna, E-mail: umeobinna@gmail.com [Department of Physics and Astronomy, University of the Western Cape, Cape Town 7535 (South Africa)
2017-06-01
Intensity mapping of the HI brightness temperature provides a unique way of tracing large-scale structures of the Universe up to the largest possible scales. This is achieved by using a low angular resolution radio telescopes to detect emission line from cosmic neutral Hydrogen in the post-reionization Universe. We use general relativistic perturbation theory techniques to derive for the first time the full expression for the HI brightness temperature up to third order in perturbation theory without making any plane-parallel approximation. We use this result and the renormalization prescription for biased tracers to study the impact of nonlinear effects on the power spectrum of HI brightness temperature both in real and redshift space. We show how mode coupling at nonlinear order due to nonlinear bias parameters and redshift space distortion terms modulate the power spectrum on large scales. The large scale modulation may be understood to be due to the effective bias parameter and effective shot noise.
The large-scale gravitational bias from the quasi-linear regime.
Bernardeau, F.
1996-08-01
It is known that in gravitational instability scenarios the nonlinear dynamics induces non-Gaussian features in cosmological density fields that can be investigated with perturbation theory. Here, I derive the expression of the joint moments of cosmological density fields taken at two different locations. The results are valid when the density fields are filtered with a top-hat filter window function, and when the distance between the two cells is large compared to the smoothing length. In particular I show that it is possible to get the generating function of the coefficients C_p,q_ defined by _c_=C_p,q_ ^p+q-2^ where δ({vec}(x)) is the local smoothed density field. It is then possible to reconstruct the joint density probability distribution function (PDF), generalizing for two points what has been obtained previously for the one-point density PDF. I discuss the validity of the large separation approximation in an explicit numerical Monte Carlo integration of the C_2,1_ parameter as a function of |{vec}(x)_1_-{vec}(x)_2_|. A straightforward application is the calculation of the large-scale ``bias'' properties of the over-dense (or under-dense) regions. The properties and the shape of the bias function are presented in details and successfully compared with numerical results obtained in an N-body simulation with CDM initial conditions.
Directory of Open Access Journals (Sweden)
Stefanie D. Hueber
2016-02-01
Full Text Available Phylogenetic methods are key to providing models for how a given protein family evolved. However, these methods run into difficulties when sequence divergence is either too low or too high. Here, we provide a case study of Hox and ParaHox proteins so that additional insights can be gained using a new computational approach to help solve old classification problems. For two (Gsx and Cdx out of three ParaHox proteins the assignments differ between the currently most established view and four alternative scenarios. We use a non-phylogenetic, pairwise-sequence-similarity-based method to assess which of the previous predictions, if any, are best supported by the sequence-similarity relationships between Hox and ParaHox proteins. The overall sequence-similarities show Gsx to be most similar to Hox2–3, and Cdx to be most similar to Hox4–8. The results indicate that a purely pairwise-sequence-similarity-based approach can provide additional information not only when phylogenetic inference methods have insufficient information to provide reliable classifications (as was shown previously for central Hox proteins, but also when the sequence variation is so high that the resulting phylogenetic reconstructions are likely plagued by long-branch-attraction artifacts.
A wideband large dynamic range and high linearity RF front-end for U-band mobile DTV
International Nuclear Information System (INIS)
Liu Rongjiang; Liu Shengyou; Guo Guiliang; Cheng Xu; Yan Yuepeng
2013-01-01
A wideband large dynamic range and high linearity U-band RF front-end for mobile DTV is introduced, and includes a noise-cancelling low-noise amplifier (LNA), an RF programmable gain amplifier (RFPGA) and a current communicating passive mixer. The noise/distortion cancelling structure and RC post-distortion compensation are employed to improve the linearity of the LNA. An RFPGA with five stages provides large dynamic range and fine gain resolution. A simple resistor voltage network in the passive mixer decreases the gate bias voltage of the mixing transistor, and optimum linearity and symmetrical mixing is obtained at the same time. The RF front-end is implemented in a 0.25 μm CMOS process. Tests show that it achieves an IIP3 (third-order intercept point) of −17 dBm, a conversion gain of 39 dB, and a noise figure of 5.8 dB. The RFPGA achieves a dynamic range of −36.2 to 23.5 dB with a resolution of 0.32 dB. (semiconductor integrated circuits)
Identifiability of large-scale non-linear dynamic network models applied to the ADM1-case study.
Nimmegeers, Philippe; Lauwers, Joost; Telen, Dries; Logist, Filip; Impe, Jan Van
2017-06-01
In this work, both the structural and practical identifiability of the Anaerobic Digestion Model no. 1 (ADM1) is investigated, which serves as a relevant case study of large non-linear dynamic network models. The structural identifiability is investigated using the probabilistic algorithm, adapted to deal with the specifics of the case study (i.e., a large-scale non-linear dynamic system of differential and algebraic equations). The practical identifiability is analyzed using a Monte Carlo parameter estimation procedure for a 'non-informative' and 'informative' experiment, which are heuristically designed. The model structure of ADM1 has been modified by replacing parameters by parameter combinations, to provide a generally locally structurally identifiable version of ADM1. This means that in an idealized theoretical situation, the parameters can be estimated accurately. Furthermore, the generally positive structural identifiability results can be explained from the large number of interconnections between the states in the network structure. This interconnectivity, however, is also observed in the parameter estimates, making uncorrelated parameter estimations in practice difficult. Copyright © 2017. Published by Elsevier Inc.
Directory of Open Access Journals (Sweden)
Aliasghar Baziar
2015-03-01
Full Text Available Abstract In order to handle large scale problems this study has used shuffled frog leaping algorithm. This algorithm is an optimization method based on natural memetics that uses a new two-phase modification to it to have a better search in the problem space. The suggested algorithm is evaluated by comparing to some well known algorithms using several benchmark optimization problems. The simulation results have clearly shown the superiority of this algorithm over other well-known methods in the area.
Energy Technology Data Exchange (ETDEWEB)
Nygaard, K
1967-12-15
The numerical deconvolution of spectra is equivalent to the solution of a (large) system of linear equations with a matrix which is not necessarily a square matrix. The demand that the square sum of the residual errors shall be minimum is not in general sufficient to ensure a unique or 'sound' solution. Therefore other demands which may include the demand for minimum square errors are introduced which lead to 'sound' and 'non-oscillatory' solutions irrespective of the shape of the original matrix and of the determinant of the matrix of the normal equations.
Experimental simulations of beam propagation over large distances in a compact linear Paul trap
International Nuclear Information System (INIS)
Gilson, Erik P.; Chung, Moses; Davidson, Ronald C.; Dorf, Mikhail; Efthimion, Philip C.; Majeski, Richard
2006-01-01
The Paul Trap Simulator Experiment (PTSX) is a compact laboratory experiment that places the physicist in the frame of reference of a long, charged-particle bunch coasting through a kilometers-long magnetic alternating-gradient (AG) transport system. The transverse dynamics of particles in both systems are described by similar equations, including nonlinear space-charge effects. The time-dependent voltages applied to the PTSX quadrupole electrodes are equivalent to the axially oscillating magnetic fields applied in the AG system. Experiments concerning the quiescent propagation of intense beams over large distances can then be performed in a compact and flexible facility. An understanding and characterization of the conditions required for quiescent beam transport, minimum halo particle generation, and precise beam compression and manipulation techniques, are essential, as accelerators and transport systems demand that ever-increasing amounts of space charge be transported. Application areas include ion-beam-driven high energy density physics, high energy and nuclear physics accelerator systems, etc. One-component cesium plasmas have been trapped in PTSX that correspond to normalized beam intensities, s=ω p 2 (0)/2ω q 2 , up to 80% of the space-charge limit where self-electric forces balance the applied focusing force. Here, ω p (0)=[n b (0)e b 2 /m b ε 0 ] 1/2 is the on-axis plasma frequency, and ω q is the smooth-focusing frequency associated with the applied focusing field. Plasmas in PTSX with values of s that are 20% of the limit have been trapped for times corresponding to equivalent beam propagation over 10 km. Results are presented for experiments in which the amplitude of the quadrupole focusing lattice is modified as a function of time. It is found that instantaneous changes in lattice amplitude can be detrimental to transverse confinement of the charge bunch
Experimental simulations of beam propagation over large distances in a compact linear Paul trapa)
Gilson, Erik P.; Chung, Moses; Davidson, Ronald C.; Dorf, Mikhail; Efthimion, Philip C.; Majeski, Richard
2006-05-01
The Paul Trap Simulator Experiment (PTSX) is a compact laboratory experiment that places the physicist in the frame of reference of a long, charged-particle bunch coasting through a kilometers-long magnetic alternating-gradient (AG) transport system. The transverse dynamics of particles in both systems are described by similar equations, including nonlinear space-charge effects. The time-dependent voltages applied to the PTSX quadrupole electrodes are equivalent to the axially oscillating magnetic fields applied in the AG system. Experiments concerning the quiescent propagation of intense beams over large distances can then be performed in a compact and flexible facility. An understanding and characterization of the conditions required for quiescent beam transport, minimum halo particle generation, and precise beam compression and manipulation techniques, are essential, as accelerators and transport systems demand that ever-increasing amounts of space charge be transported. Application areas include ion-beam-driven high energy density physics, high energy and nuclear physics accelerator systems, etc. One-component cesium plasmas have been trapped in PTSX that correspond to normalized beam intensities, ŝ=ωp2(0)/2ωq2, up to 80% of the space-charge limit where self-electric forces balance the applied focusing force. Here, ωp(0)=[nb(0)eb2/mbɛ0]1/2 is the on-axis plasma frequency, and ωq is the smooth-focusing frequency associated with the applied focusing field. Plasmas in PTSX with values of ŝ that are 20% of the limit have been trapped for times corresponding to equivalent beam propagation over 10km. Results are presented for experiments in which the amplitude of the quadrupole focusing lattice is modified as a function of time. It is found that instantaneous changes in lattice amplitude can be detrimental to transverse confinement of the charge bunch.
International Nuclear Information System (INIS)
Bhattacharya, Deb Sankar; Majumdar, Nayana; Sarkar, S.; Bhattacharya, S.; Mukhopadhyay, Supratik; Bhattacharya, P.; Attie, D.; Colas, P.; Ganjour, S.; Bhattacharya, Aparajita
2016-01-01
The principal particle tracker at the International Linear Collider (ILC) is planned to be a large Time Projection Chamber (TPC) where different Micro Pattern Gaseous Detector (MPGDs) candidate as the gaseous amplifier. A Micromegas (MM) based TPC can meet the ILC requirement of continuous and precise pattern recognition. Seven MM modules, working as the end-plate of a Large Prototype TPC (LPTPC) installed at DESY, have been tested with a 5 GeV electron beam. Due to the grounded peripheral frame of the MM modules, at low drift, the electric field lines near the detector edge remain no longer parallel to the TPC axis. This causes signal loss along the boundaries of the MM modules as well as distortion in the reconstructed track. In presence of magnetic field, the distorted electric field introduces ExB effect
Hargrove, W. W.; Hoffman, F. M.; Kumar, J.; Spruce, J.; Norman, S. P.
2013-12-01
Here we present diverse examples where empirical mining and statistical analysis of large data sets have already been shown to be useful for a wide variety of practical decision-making problems within the realm of large-scale ecology. Because a full understanding and appreciation of particular ecological phenomena are possible only after hypothesis-directed research regarding the existence and nature of that process, some ecologists may feel that purely empirical data harvesting may represent a less-than-satisfactory approach. Restricting ourselves exclusively to process-driven approaches, however, may actually slow progress, particularly for more complex or subtle ecological processes. We may not be able to afford the delays caused by such directed approaches. Rather than attempting to formulate and ask every relevant question correctly, empirical methods allow trends, relationships and associations to emerge freely from the data themselves, unencumbered by a priori theories, ideas and prejudices that have been imposed upon them. Although they cannot directly demonstrate causality, empirical methods can be extremely efficient at uncovering strong correlations with intermediate "linking" variables. In practice, these correlative structures and linking variables, once identified, may provide sufficient predictive power to be useful themselves. Such correlation "shadows" of causation can be harnessed by, e.g., Bayesian Belief Nets, which bias ecological management decisions, made with incomplete information, toward favorable outcomes. Empirical data-harvesting also generates a myriad of testable hypotheses regarding processes, some of which may even be correct. Quantitative statistical regionalizations based on quantitative multivariate similarity have lended insights into carbon eddy-flux direction and magnitude, wildfire biophysical conditions, phenological ecoregions useful for vegetation type mapping and monitoring, forest disease risk maps (e.g., sudden oak
Minnett, R.; Koppers, A. A.; Tauxe, L.; Constable, C.; Jarboe, N. A.
2011-12-01
The Magnetics Information Consortium (MagIC) provides an archive for the wealth of rock- and paleomagnetic data and interpretations from studies on natural and synthetic samples. As with many fields, most peer-reviewed paleo- and rock magnetic publications only include high level results. However, access to the raw data from which these results were derived is critical for compilation studies and when updating results based on new interpretation and analysis methods. MagIC provides a detailed metadata model with places for everything from raw measurements to their interpretations. Prior to MagIC, these raw data were extremely cumbersome to collect because they mostly existed in a lab's proprietary format on investigator's personal computers or undigitized in field notebooks. MagIC has developed a suite of offline and online tools to enable the paleomagnetic, rock magnetic, and affiliated scientific communities to easily contribute both their previously published data and data supporting an article undergoing peer-review, to retrieve well-annotated published interpretations and raw data, and to analyze and visualize large collections of published data online. Here we present the technology we chose (including VBA in Excel spreadsheets, Python libraries, FastCGI JSON webservices, Oracle procedures, and jQuery user interfaces) and how we implemented it in order to serve the scientific community as seamlessly as possible. These tools are now in use in labs worldwide, have helped archive many valuable legacy studies and datasets, and routinely enable new contributions to the MagIC Database (http://earthref.org/MAGIC/).
Fast Solvers for Dense Linear Systems
Energy Technology Data Exchange (ETDEWEB)
Kauers, Manuel [Research Institute for Symbolic Computation (RISC), Altenbergerstrasse 69, A4040 Linz (Austria)
2008-10-15
It appears that large scale calculations in particle physics often require to solve systems of linear equations with rational number coefficients exactly. If classical Gaussian elimination is applied to a dense system, the time needed to solve such a system grows exponentially in the size of the system. In this tutorial paper, we present a standard technique from computer algebra that avoids this exponential growth: homomorphic images. Using this technique, big dense linear systems can be solved in a much more reasonable time than using Gaussian elimination over the rationals.
Directory of Open Access Journals (Sweden)
Maria S. Prokhorova
2014-01-01
Full Text Available The article deals with a study of problemsof ﬁnding the optimal portfolio securitiesusing convolutions expectation of portfolioreturns and portfolio variance. Value of thecoefﬁcient of risk, in which the problem ofmaximizing the variance - limited yieldis equivalent to maximizing a linear convolution of criteria for «expected returns-variance» is obtained. An automated method for ﬁnding the optimal portfolio, onthe basis of which the results of the studydemonstrated is proposed.
Lee, S. H.; Shulika, Olga.; Kim, K. N.; Yeom, G. Y.; Lee, J. K.
2004-09-01
As the technology of plasma processing progresses, there is a continuing demand for higher plasma density, uniformity over large areas and greater control over plasma parameters to optimize the processes of etching, deposition and surface treatment. Traditionally, the external planar ICP sources with low pressure high density plasma have limited scale-up capabilities due to its high impedance accompanied by the large antenna size. Also due to the cost and thickness of their dielectric material in order to generate uniform plasma. In this study the novel internal-type linear inductive antenna system (1,020mm¡¿830mm¡¿437mm) with permanent magnet arrays are investigated to improve both the plasma density and the uniformity of LAPS (Large Area Plasma Source) for FPD processing. Generally plasma discharges are enhanced because the inductance of the novel antenna (termed as the double comb antenna) is lower than that of the serpentine-type antenna and also the magnetic confinement of electron increases the power absorption efficiency. The uniformity is improved by reducing the standing wave effect. The total length of antenna is comparable to the driving rf wavelength to cause the plasma nonuniformity. To describe the discharge phenomenon we have developed a magnetized two-dimensional fluid simulation. This work was supported by National Research Laboratory (NRL) Program of the Korea Ministry of Science and Technology. [References] 1. J.K.Lee, Lin Meng, Y.K.Shin, H,J,Lee and T.H.Chung, ¡°Modeling and Simulation of a Large-Area Plasma Source¡±, Jpn. J. Appl. Phys. Vol.36(1997) pp. 5714-5723 2. S.E.Park, B.U.Cho, Y.J.Lee*, and G.Y.Yeom*, and J.K.Lee, ¡°The Characteristics of Large Area Processing Plasmas¡±, IEEE Trans. Plasma Sci., Vol.31 ,No.4(2003) pp. 628-637
Sagar, Rizwan Ur Rehman; Galluzzi, Massimiliano; Wan, Caihua; Shehzad, Khurram; Navale, Sachin T; Anwar, Tauseef; Mane, Rajaram S; Piao, Hong-Guang; Ali, Abid; Stadler, Florian J
2017-01-18
Here, we present the first observation of magneto-transport properties of graphene foam (GF) composed of a few layers in a wide temperature range of 2-300 K. Large room-temperature linear positive magnetoresistance (PMR ≈ 171% at B ≈ 9 T) has been detected. The largest PMR (∼213%) has been achieved at 2 K under a magnetic field of 9 T, which can be tuned by the addition of poly(methyl methacrylate) to the porous structure of the foam. This remarkable magnetoresistance may be the result of quadratic magnetoresistance. The excellent magneto-transport properties of GF open a way toward three-dimensional graphene-based magnetoelectronic devices.
A novel large thrust-weight ratio V-shaped linear ultrasonic motor with a flexible joint.
Li, Xiaoniu; Yao, Zhiyuan; Yang, Mojian
2017-06-01
A novel large thrust-weight ratio V-shaped linear ultrasonic motor with a flexible joint is proposed in this paper. The motor is comprised of a V-shaped transducer, a slider, a clamp, and a base. The V-shaped transducer consists of two piezoelectric beams connected through a flexible joint to form an appropriate coupling angle. The V-shaped motor is operated in the coupled longitudinal-bending mode. Longitudinal and bending movements are transferred by the flexible joint between the two beams. Compared with the coupled longitudinal-bending mode of the single piezoelectric beam or the symmetrical and asymmetrical modes of the previous V-shaped transducer, the coupled longitudinal-bending mode of the V-shaped transducer with a flexible joint provides higher vibration efficiency and more convenient mode conformance adjustment. A finite element model of the V-shaped transducer is created to numerically study the influence of geometrical parameters and to determine the final geometrical parameters. In this paper, three prototypes were then fabricated and experimentally investigated. The modal test results match well with the finite element analysis. The motor mechanical output characteristics of three different coupling angles θ indicate that V-90 (θ = 90°) is the optimal angle. The mechanical output experiments conducted using the V-90 prototype (Size: 59.4 mm × 30.7 mm × 4 mm) demonstrate that the maximum unloaded speed is 1.2 m/s under a voltage of 350 Vpp, and the maximum output force is 15 N under a voltage of 300 Vpp. The proposed novel V-shaped linear ultrasonic motor has a compact size and a simple structure with a large thrust-weight ratio (0.75 N/g) and high speed.
Green, David L.; Berry, Lee A.; Simpson, Adam B.; Younkin, Timothy R.
2018-04-01
We present the KINETIC-J code, a computational kernel for evaluating the linearized Vlasov equation with application to calculating the kinetic plasma response (current) to an applied time harmonic wave electric field. This code addresses the need for a configuration space evaluation of the plasma current to enable kinetic full-wave solvers for waves in hot plasmas to move beyond the limitations of the traditional Fourier spectral methods. We benchmark the kernel via comparison with the standard k →-space forms of the hot plasma conductivity tensor.
International Nuclear Information System (INIS)
Martins, Cesar C.; Bicego, Marcia C.; Mahiques, Michel M.; Figueira, Rubens C.L.; Tessler, Moyses G.; Montone, Rosalinda C.
2010-01-01
This paper reports the reconstruction of the contamination history of a large South American industrial coastal area (Santos Estuary, Brazil) using linear alkylbenzenes (LABs). Three sediment cores were dated by 137 Cs. Concentrations in surficial layers were comparable to the midrange concentrations reported for coastal sediments worldwide. LAB concentrations increased towards the surface, indicating increased waste discharges into the estuary in recent decades. The highest concentration values occurred in the early 1970s, a time of intense industrial activity and marked population growth. The decreased LAB concentration, in the late 1970s was assumed to be the result of the world oil crisis. Treatment of industrial effluents, which began in 1984, was represented by decreased LAB levels. Microbial degradation of LABs may be more intense in the industrial area sediments. The results show that industrial and domestic waste discharges are a historical problem in the area. - The contamination history of a large South American industrial coastal area indicated by molecular indicator of sewage input.
Dong, Peng; Pérez-Andújar, Angélica; Pinnaduwage, Dilini; Braunstein, Steve; Theodosopoulos, Philip; McDermott, Michael; Sneed, Penny; Ma, Lijun
2016-12-01
OBJECTIVE Noninvasive Gamma Knife (GK) platforms, such as the relocatable frame and on-board imaging, have enabled hypofractionated GK radiosurgery of large or complex brain lesions. This study aimed to characterize the dosimetric quality of such treatments against linear accelerator-based delivery systems that include the CyberKnife (CK) and volumetric modulated arc therapy (VMAT). METHODS Ten patients treated with VMAT at the authors' institution for large brain tumors (> 3 cm in maximum diameter) were selected for the study. The median prescription dose was 25 Gy (range 20-30 Gy) in 5 fractions. The median planning target volume (PTV) was 9.57 cm 3 (range 1.94-24.81 cm 3 ). Treatment planning was performed using Eclipse External Beam Planning V11 for VMAT on the Varian TrueBeam system, Multiplan V4.5 for the CyberKnife VSI System, and Leksell GammaPlan V10.2 for the Gamma Knife Perfexion system. The percentage of the PTV receiving at least the prescription dose was normalized to be identical across all platforms for individual cases. The prescription isodose value for the PTV, conformity index, Paddick gradient index, mean and maximum doses for organs at risk, and normal brain dose at variable isodose volumes ranging from the 5-Gy isodose volume (V5) to the 15-Gy isodose volume (V15) were compared for all of the cases. RESULTS The mean Paddick gradient index was 2.6 ± 0.2, 3.2 ± 0.5, and 4.3 ± 1.0 for GK, CK, and VMAT, respectively (p 0.06). The average prescription isodose values were 52% (range 47%-69%), 60% (range 46%-68%), and 88% (range 70%-94%) for GK, CK, and VMAT, respectively, thus producing significant variations in dose hot spots among the 3 platforms. Furthermore, the mean V5 values for GK and CK were similar (p > 0.79) at 71.9 ± 36.2 cm 3 and 73.3 ± 31.8 cm 3 , respectively, both of which were statistically lower (p linear accelerator-based treatments. Such a result supports the use of a large number of isocenters or confocal beams for the
Amirghasemi, Mehrdad; Zamani, Reza
2014-01-01
This paper presents an effective procedure for solving the job shop problem. Synergistically combining small and large neighborhood schemes, the procedure consists of four components, namely (i) a construction method for generating semi-active schedules by a forward-backward mechanism, (ii) a local search for manipulating a small neighborhood structure guided by a tabu list, (iii) a feedback-based mechanism for perturbing the solutions generated, and (iv) a very large-neighborhood local search guided by a forward-backward shifting bottleneck method. The combination of shifting bottleneck mechanism and tabu list is used as a means of the manipulation of neighborhood structures, and the perturbation mechanism employed diversifies the search. A feedback mechanism, called repeat-check, detects consequent repeats and ignites a perturbation when the total number of consecutive repeats for two identical makespan values reaches a given threshold. The results of extensive computational experiments on the benchmark instances indicate that the combination of these four components is synergetic, in the sense that they collectively make the procedure fast and robust.
Tuey, R. C.
1972-01-01
Computer solutions of linear programming problems are outlined. Information covers vector spaces, convex sets, and matrix algebra elements for solving simultaneous linear equations. Dual problems, reduced cost analysis, ranges, and error analysis are illustrated.
Nowhere to hide: Effects of linear features on predator-prey dynamics in a large mammal system.
DeMars, Craig A; Boutin, Stan
2018-01-01
Rapid landscape alteration associated with human activity is currently challenging the evolved dynamical stability of many predator-prey systems by forcing species to behaviourally respond to novel environmental stimuli. In many forested systems, linear features (LFs) such as roads, pipelines and resource exploration lines (i.e. seismic lines) are a ubiquitous form of landscape alteration that have been implicated in altering predator-prey dynamics. One hypothesized effect is that LFs facilitate predator movement into and within prey refugia, thereby increasing predator-prey spatial overlap. We evaluated this hypothesis in a large mammal system, focusing on the interactions between boreal woodland caribou (Rangifer tarandus caribou) and their two main predators, wolves (Canis lupus) and black bears (Ursus americanus), during the calving season of caribou. In this system, LFs extend into and occur within peatlands (i.e. bogs and nutrient-poor fens), a habitat type highly used by caribou due to its refugia effects. Using resource selection analyses, we found that LFs increased predator selection of peatlands. Female caribou appeared to respond by avoiding LFs and areas with high LF density. However, in our study area, most caribou cannot completely avoid exposure to LFs and variation in female response had demographic effects. In particular, increasing proportional use of LFs by females negatively impacted survival of their neonate calves. Collectively, these results demonstrate how LFs can reduce the efficacy of prey refugia. Mitigating such effects will require limiting or restoring LFs within prey refugia, although the effectiveness of mitigation efforts will depend upon spatial scale, which in turn will be influenced by the life-history traits of predator and prey. © 2017 The Authors. Journal of Animal Ecology © 2017 British Ecological Society.
International Nuclear Information System (INIS)
Xiao, Gang; Jia, Ming; Wang, Tianyou
2016-01-01
Spray combustion of n-heptane in a constant-volume vessel under engine-relevant conditions was investigated using linear eddy model in the framework of large eddy simulation. In this numerical approach, turbulent mixing was traced by an innovative stochastic approach instead of the conventional gradient diffusion model. Chemical reaction rates were calculated with the consideration of the sub-grid scale spatial fluctuations of reactive scalars. Turbulence-chemistry interactions were represented by the separated treatments of the underlying processes including turbulent stirring, chemical reaction, and molecular diffusion. The model was validated against the experimental data of ignition delay times, chemiluminescence images, and soot images from Sandia National Laboratories. Numerical results showed that the ignition process changed from the temperature-controlled regime to the mixing-controlled regime as the initial ambient temperature increased from 800 K to 1000 K. The premixed flame and the diffusion flame coexisted, while the gross heat release rate was found to be dominated by the premixed flame. The temperature fluctuation was mainly observed around the spray jet due to the cooling effect of the fuel vaporization. The fluctuations were more significantly smoothed out by the high-temperature flame than the low-temperature flame. The mean temperature would be overpredicted if the sub-grid temperature fluctuation was neglected. - Highlights: • Turbulent mixing is traced by stochastic method instead of gradient diffusion model. • Sub-grid scale fluctuations of reactive scalars are captured. • Ignition process varies from temperature-controlled to mixing-controlled regime. • Temperature fluctuation can be smoothed out by high-temperature flame. • The heat release rate is dominated by the premixed flame.
Monte Carlo method for solving a parabolic problem
Directory of Open Access Journals (Sweden)
Tian Yi
2016-01-01
Full Text Available In this paper, we present a numerical method based on random sampling for a parabolic problem. This method combines use of the Crank-Nicolson method and Monte Carlo method. In the numerical algorithm, we first discretize governing equations by Crank-Nicolson method, and obtain a large sparse system of linear algebraic equations, then use Monte Carlo method to solve the linear algebraic equations. To illustrate the usefulness of this technique, we apply it to some test problems.
Wei, Haiqiao; Zhao, Wanhui; Zhou, Lei; Chen, Ceyuan; Shu, Gequn
2018-03-01
Large eddy simulation coupled with the linear eddy model (LEM) is employed for the simulation of n-heptane spray flames to investigate the low temperature ignition and combustion process in a constant-volume combustion vessel under diesel-engine relevant conditions. Parametric studies are performed to give a comprehensive understanding of the ignition processes. The non-reacting case is firstly carried out to validate the present model by comparing the predicted results with the experimental data from the Engine Combustion Network (ECN). Good agreements are observed in terms of liquid and vapour penetration length, as well as the mixture fraction distributions at different times and different axial locations. For the reacting cases, the flame index was introduced to distinguish between the premixed and non-premixed combustion. A reaction region (RR) parameter is used to investigate the ignition and combustion characteristics, and to distinguish the different combustion stages. Results show that the two-stage combustion process can be identified in spray flames, and different ignition positions in the mixture fraction versus RR space are well described at low and high initial ambient temperatures. At an initial condition of 850 K, the first-stage ignition is initiated at the fuel-lean region, followed by the reactions in fuel-rich regions. Then high-temperature reaction occurs mainly at the places with mixture concentration around stoichiometric mixture fraction. While at an initial temperature of 1000 K, the first-stage ignition occurs at the fuel-rich region first, then it moves towards fuel-richer region. Afterwards, the high-temperature reactions move back to the stoichiometric mixture fraction region. For all of the initial temperatures considered, high-temperature ignition kernels are initiated at the regions richer than stoichiometric mixture fraction. By increasing the initial ambient temperature, the high-temperature ignition kernels move towards richer
Fuzzy linear programming approach for solving transportation ...
Indian Academy of Sciences (India)
ALI EBRAHIMNEJAD
Department of Mathematics, Qaemshahr Branch, Islamic Azad University, Qaemshahr, Iran e-mail: ..... est grade of membership at x are μ ˜AL (x) and μ ˜AU (x), respectively. ..... trapezoidal fuzzy numbers transportation problem (12) are.
ITMETH, Iterative Routines for Linear System
International Nuclear Information System (INIS)
Greenbaum, A.
1989-01-01
1 - Description of program or function: ITMETH is a collection of iterative routines for solving large, sparse linear systems. 2 - Method of solution: ITMETH solves general linear systems of the form AX=B using a variety of methods: Jacobi iteration; Gauss-Seidel iteration; incomplete LU decomposition or matrix splitting with iterative refinement; diagonal scaling, matrix splitting, or incomplete LU decomposition with the conjugate gradient method for the problem AA'Y=B, X=A'Y; bi-conjugate gradient method with diagonal scaling, matrix splitting, or incomplete LU decomposition; and ortho-min method with diagonal scaling, matrix splitting, or incomplete LU decomposition. ITMETH also solves symmetric positive definite linear systems AX=B using the conjugate gradient method with diagonal scaling or matrix splitting, or the incomplete Cholesky conjugate gradient method
Bai, Y.Q.; Lesaja, G.; Roos, C.; Wang, G.Q.; El Ghami, M.
2008-01-01
In this paper we present a class of polynomial primal-dual interior-point algorithms for linear optimization based on a new class of kernel functions. This class is fairly general and includes the classical logarithmic function, the prototype self-regular function, and non-self-regular kernel
Large-scale matrix-handling subroutines 'ATLAS'
International Nuclear Information System (INIS)
Tsunematsu, Toshihide; Takeda, Tatsuoki; Fujita, Keiichi; Matsuura, Toshihiko; Tahara, Nobuo
1978-03-01
Subroutine package ''ATLAS'' has been developed for handling large-scale matrices. The package is composed of four kinds of subroutines, i.e., basic arithmetic routines, routines for solving linear simultaneous equations and for solving general eigenvalue problems and utility routines. The subroutines are useful in large scale plasma-fluid simulations. (auth.)
Edwards, Harold M
1995-01-01
In his new undergraduate textbook, Harold M Edwards proposes a radically new and thoroughly algorithmic approach to linear algebra Originally inspired by the constructive philosophy of mathematics championed in the 19th century by Leopold Kronecker, the approach is well suited to students in the computer-dominated late 20th century Each proof is an algorithm described in English that can be translated into the computer language the class is using and put to work solving problems and generating new examples, making the study of linear algebra a truly interactive experience Designed for a one-semester course, this text adopts an algorithmic approach to linear algebra giving the student many examples to work through and copious exercises to test their skills and extend their knowledge of the subject Students at all levels will find much interactive instruction in this text while teachers will find stimulating examples and methods of approach to the subject
Nasari, Masoud M; Szyszkowicz, Mieczysław; Chen, Hong; Crouse, Daniel; Turner, Michelle C; Jerrett, Michael; Pope, C Arden; Hubbell, Bryan; Fann, Neal; Cohen, Aaron; Gapstur, Susan M; Diver, W Ryan; Stieb, David; Forouzanfar, Mohammad H; Kim, Sun-Young; Olives, Casey; Krewski, Daniel; Burnett, Richard T
2016-01-01
The effectiveness of regulatory actions designed to improve air quality is often assessed by predicting changes in public health resulting from their implementation. Risk of premature mortality from long-term exposure to ambient air pollution is the single most important contributor to such assessments and is estimated from observational studies generally assuming a log-linear, no-threshold association between ambient concentrations and death. There has been only limited assessment of this assumption in part because of a lack of methods to estimate the shape of the exposure-response function in very large study populations. In this paper, we propose a new class of variable coefficient risk functions capable of capturing a variety of potentially non-linear associations which are suitable for health impact assessment. We construct the class by defining transformations of concentration as the product of either a linear or log-linear function of concentration multiplied by a logistic weighting function. These risk functions can be estimated using hazard regression survival models with currently available computer software and can accommodate large population-based cohorts which are increasingly being used for this purpose. We illustrate our modeling approach with two large cohort studies of long-term concentrations of ambient air pollution and mortality: the American Cancer Society Cancer Prevention Study II (CPS II) cohort and the Canadian Census Health and Environment Cohort (CanCHEC). We then estimate the number of deaths attributable to changes in fine particulate matter concentrations over the 2000 to 2010 time period in both Canada and the USA using both linear and non-linear hazard function models.
A scalable parallel algorithm for multiple objective linear programs
Wiecek, Malgorzata M.; Zhang, Hong
1994-01-01
This paper presents an ADBASE-based parallel algorithm for solving multiple objective linear programs (MOLP's). Job balance, speedup and scalability are of primary interest in evaluating efficiency of the new algorithm. Implementation results on Intel iPSC/2 and Paragon multiprocessors show that the algorithm significantly speeds up the process of solving MOLP's, which is understood as generating all or some efficient extreme points and unbounded efficient edges. The algorithm gives specially good results for large and very large problems. Motivation and justification for solving such large MOLP's are also included.
Directory of Open Access Journals (Sweden)
Aleksa S. Srdanov
2018-04-01
Full Text Available Solving a linear system of n × n equations can be very difficult for the computer, especially if one needs the exact solution, even when the number n - of equations and of unknown variables is relatively small (a few thousands. All existing methods have to overcome at least one of the following problems: 1. Computational complexity, which is expressed with the number of arithmetic operations required in order to determine a solution; 2. The possibility of overflow and underflow problems; 3. Causing variations in the values of some coefficients in the initial system, which may be leading to instability of the solution; 4. Requiring additional conditions for convergence; 5. In cases of a large number of equations and unknown variables it is often required that the systems matrix be: either sparse, or symmetrical, or diagonal, etc. This paper presents a method for solving a system of linear equations of arbitrary order (any number of equations and unknown variables to which the problems listed above do not reflect. / Решение систем линейных уравнений n × n может представлять проблему для компьютера, особенно в тех случаях, когда требуется точное решение, и даже в тех случаях, когда количество уравнений и неизвестных относительно невелико (всего несколько тысяч. Все существующие методы сталкиваются с наименее одной из следующего ряда проблем: 1. сложность вычисления, выраженная количеством соответствующих операций, которые необходимо произвести для получения решения; 2. потенциальная возможность неограниченного роста
DEFF Research Database (Denmark)
Ritz, Christian; Laursen, Rikke Pilmann; Damsgaard, Camilla Trab
2017-01-01
of a school meal programme. We propose a novel and versatile framework for simultaneous inference on parameters estimated from linear mixed models that were fitted separately for several outcomes from the same study, but did not necessarily contain the same fixed or random effects. By combining asymptotic...... sizes of practical relevance we studied simultaneous coverage through simulation, which showed that the approach achieved acceptable coverage probabilities even for small sample sizes (10 clusters) and for 2–16 outcomes. The approach also compared favourably with a joint modelling approach. We also...
Pavlosiuk, Orest; Kaczorowski, Dariusz; Wiśniewski, Piotr
2015-01-01
We present electronic transport and magnetic properties of single crystals of semimetallic half-Heusler phase LuPdBi, having theoretically predicted band inversion requisite for nontrivial topological properties. The compound exhibits superconductivity below a critical temperature Tc = 1.8 K, with a zero-temperature upper critical field Bc2 ≈ 2.3 T. Although superconducting state is clearly reflected in the electrical resistivity and magnetic susceptibility data, no corresponding anomaly can be seen in the specific heat. Temperature dependence of the electrical resistivity suggests existence of two parallel conduction channels: metallic and semiconducting, with the latter making negligible contribution at low temperatures. The magnetoresistance is huge and clearly shows a weak antilocalization effect in small magnetic fields. Above about 1.5 T, the magnetoresistance becomes linear and does not saturate in fields up to 9 T. The linear magnetoresistance is observed up to room temperature. Below 10 K, it is accompanied by Shubnikov-de Haas oscillations. Their analysis reveals charge carriers with effective mass of 0.06 me and a Berry phase very close to π, expected for Dirac-fermion surface states, thus corroborating topological nature of the material. PMID:25778789
Solving Environmental Problems
DEFF Research Database (Denmark)
Ørding Olsen, Anders; Sofka, Wolfgang; Grimpe, Christoph
2017-01-01
for Research and Technological Development (FP7), our results indicate that the problem-solving potential of a search strategy increases with the diversity of existing knowledge of the partners in a consortium and with the experience of the partners involved. Moreover, we identify a substantial negative effect...... dispersed. Hence, firms need to collaborate. We shed new light on collaborative search strategies led by firms in general and for solving environmental problems in particular. Both topics are largely absent in the extant open innovation literature. Using data from the European Seventh Framework Program...
Transport equation solving methods
International Nuclear Information System (INIS)
Granjean, P.M.
1984-06-01
This work is mainly devoted to Csub(N) and Fsub(N) methods. CN method: starting from a lemma stated by Placzek, an equivalence is established between two problems: the first one is defined in a finite medium bounded by a surface S, the second one is defined in the whole space. In the first problem the angular flux on the surface S is shown to be the solution of an integral equation. This equation is solved by Galerkin's method. The Csub(N) method is applied here to one-velocity problems: in plane geometry, slab albedo and transmission with Rayleigh scattering, calculation of the extrapolation length; in cylindrical geometry, albedo and extrapolation length calculation with linear scattering. Fsub(N) method: the basic integral transport equation of the Csub(N) method is integrated on Case's elementary distributions; another integral transport equation is obtained: this equation is solved by a collocation method. The plane problems solved by the Csub(N) method are also solved by the Fsub(N) method. The Fsub(N) method is extended to any polynomial scattering law. Some simple spherical problems are also studied. Chandrasekhar's method, collision probability method, Case's method are presented for comparison with Csub(N) and Fsub(N) methods. This comparison shows the respective advantages of the two methods: a) fast convergence and possible extension to various geometries for Csub(N) method; b) easy calculations and easy extension to polynomial scattering for Fsub(N) method [fr
A feasible DY conjugate gradient method for linear equality constraints
LI, Can
2017-09-01
In this paper, we propose a feasible conjugate gradient method for solving linear equality constrained optimization problem. The method is an extension of the Dai-Yuan conjugate gradient method proposed by Dai and Yuan to linear equality constrained optimization problem. It can be applied to solve large linear equality constrained problem due to lower storage requirement. An attractive property of the method is that the generated direction is always feasible and descent direction. Under mild conditions, the global convergence of the proposed method with exact line search is established. Numerical experiments are also given which show the efficiency of the method.
Systems of Inhomogeneous Linear Equations
Scherer, Philipp O. J.
Many problems in physics and especially computational physics involve systems of linear equations which arise e.g. from linearization of a general nonlinear problem or from discretization of differential equations. If the dimension of the system is not too large standard methods like Gaussian elimination or QR decomposition are sufficient. Systems with a tridiagonal matrix are important for cubic spline interpolation and numerical second derivatives. They can be solved very efficiently with a specialized Gaussian elimination method. Practical applications often involve very large dimensions and require iterative methods. Convergence of Jacobi and Gauss-Seidel methods is slow and can be improved by relaxation or over-relaxation. An alternative for large systems is the method of conjugate gradients.
Wang, Fei; Chen, Quanjiao; Li, Shuntang; Zhang, Chenyao; Li, Shanshan; Liu, Min; Mei, Kun; Li, Chunhua; Ma, Lixin; Yu, Xiaolan
2017-06-01
Linear DNA vaccines provide effective vaccination. However, their application is limited by high cost and small scale of the conventional polymerase chain reaction (PCR) generally used to obtain sufficient amounts of DNA effective against epidemic diseases. In this study, a two-step, large-scale PCR was established using a low-cost DNA polymerase, RKOD, expressed in Pichia pastoris. Two linear DNA vaccines encoding influenza H1N1 hemagglutinin (HA) 1, LEC-HA, and PTO-LEC-HA (with phosphorothioate-modified primers), were produced by the two-step PCR. Protective effects of the vaccines were evaluated in a mouse model. BALB/c mice were immunized three times with the vaccines or a control DNA fragment. All immunized animals were challenged by intranasal administration of a lethal dose of influenza H1N1 virus 2 weeks after the last immunization. Sera of the immunized animals were tested for the presence of HA-specific antibodies, and the total IFN-γ responses induced by linear DNA vaccines were measured. The results showed that the DNA vaccines but not the control DNA induced strong antibody and IFN-γ responses. Additionally, the PTO-LEC-HA vaccine effectively protected the mice against the lethal homologous mouse-adapted virus, with a survival rate of 100% versus 70% in the LEC-HA-vaccinated group, showing that the PTO-LEC-HA vaccine was more effective than LEC-HA. In conclusion, the results indicated that the linear H1N1 HA-coding DNA vaccines induced significant immune responses and protected mice against a lethal virus challenge. Thus, the low-cost, two-step, large-scale PCR can be considered a potential tool for rapid manufacturing of linear DNA vaccines against emerging infectious diseases. Copyright © 2017 Elsevier B.V. All rights reserved.
Molina, J. M.; Zaitchik, B. F.
2016-12-01
Recent findings considering high CO2 emission scenarios (RCP8.5) suggest that the tropical Andes may experience a massive warming and a significant precipitation increase (decrease) during the wet (dry) seasons by the end of the 21st century. Variations on rainfall-streamflow relationships and seasonal crop yields significantly affect human development in this region and make local communities highly vulnerable to climate change and variability. We developed an expert-informed empirical statistical downscaling (ESD) algorithm to explore and construct robust global climate predictors to perform skillful RCP8.5 projections of in-situ March-May (MAM) precipitation required for impact modeling and adaptation studies. We applied our framework to a topographically-complex region of the Colombian Andes where a number of previous studies have reported El Niño-Southern Oscillation (ENSO) as the main driver of climate variability. Supervised machine learning algorithms were trained with customized and bias-corrected predictors from NCEP reanalysis, and a cross-validation approach was implemented to assess both predictive skill and model selection. We found weak and not significant teleconnections between precipitation and lagged seasonal surface temperatures over El Niño3.4 domain, which suggests that ENSO fails to explain MAM rainfall variability in the study region. In contrast, series of Sea Level Pressure (SLP) over American Samoa -likely associated with the South Pacific Convergence Zone (SPCZ)- explains more than 65% of the precipitation variance. The best prediction skill was obtained with Selected Generalized Additive Models (SGAM) given their ability to capture linear/nonlinear relationships present in the data. While SPCZ-related series exhibited a positive linear effect in the rainfall response, SLP predictors in the north Atlantic and central equatorial Pacific showed nonlinear effects. A multimodel (MIROC, CanESM2 and CCSM) ensemble of ESD projections revealed
Kang, Bongmun; Yoon, Ho-Sung
2015-02-01
Recently, microalgae was considered as a renewable energy for fuel production because its production is nonseasonal and may take place on nonarable land. Despite all of these advantages, microalgal oil production is significantly affected by environmental factors. Furthermore, the large variability remains an important problem in measurement of algae productivity and compositional analysis, especially, the total lipid content. Thus, there is considerable interest in accurate determination of total lipid content during the biotechnological process. For these reason, various high-throughput technologies were suggested for accurate measurement of total lipids contained in the microorganisms, especially oleaginous microalgae. In addition, more advanced technologies were employed to quantify the total lipids of the microalgae without a pretreatment. However, these methods are difficult to measure total lipid content in wet form microalgae obtained from large-scale production. In present study, the thermal analysis performed with two-step linear temeperature program was applied to measure heat evolved in temperature range from 310 to 351 °C of Nostoc sp. KNUA003 obtained from large-scale cultivation. And then, we examined the relationship between the heat evolved in 310-351 °C (HE) and total lipid content of the wet Nostoc cell cultivated in raceway. As a result, the linear relationship was determined between HE value and total lipid content of Nostoc sp. KNUA003. Particularly, there was a linear relationship of 98% between the HE value and the total lipid content of the tested microorganism. Based on this relationship, the total lipid content converted from the heat evolved of wet Nostoc sp. KNUA003 could be used for monitoring its lipid induction in large-scale cultivation. Copyright © 2014 Elsevier Inc. All rights reserved.
DEFF Research Database (Denmark)
Jacobsen, Martin; Martinussen, Torben
2016-01-01
Pseudo-values have proven very useful in censored data analysis in complex settings such as multi-state models. It was originally suggested by Andersen et al., Biometrika, 90, 2003, 335 who also suggested to estimate standard errors using classical generalized estimating equation results. These r......Pseudo-values have proven very useful in censored data analysis in complex settings such as multi-state models. It was originally suggested by Andersen et al., Biometrika, 90, 2003, 335 who also suggested to estimate standard errors using classical generalized estimating equation results....... These results were studied more formally in Graw et al., Lifetime Data Anal., 15, 2009, 241 that derived some key results based on a second-order von Mises expansion. However, results concerning large sample properties of estimates based on regression models for pseudo-values still seem unclear. In this paper......, we study these large sample properties in the simple setting of survival probabilities and show that the estimating function can be written as a U-statistic of second order giving rise to an additional term that does not vanish asymptotically. We further show that previously advocated standard error...
Donnelly, Lane F; Basta, Kathryne C; Dykes, Anne M; Zhang, Wei; Shook, Joan E
2018-01-01
At a pediatric health system, the Daily Operational Brief (DOB) was updated in 2015 after three years of operation. Quality and safety metrics, the patient volume and staffing assessment, and the readiness assessment are all presented. In addition, in the problem-solving accountability system, problematic issues are categorized as Quick Hits or Complex Issues. Walk-the-Wall, a biweekly meeting attended by hospital senior administrative leadership and quality and safety leaders, is conducted to chart current progress on Complex Issues. The DOB provides a daily standardized approach to evaluate readiness to provide care to current patients and improvement in the care to be provided for future patients. Copyright © 2017 The Joint Commission. Published by Elsevier Inc. All rights reserved.
Linear programming foundations and extensions
Vanderbei, Robert J
2001-01-01
Linear Programming: Foundations and Extensions is an introduction to the field of optimization. The book emphasizes constrained optimization, beginning with a substantial treatment of linear programming, and proceeding to convex analysis, network flows, integer programming, quadratic programming, and convex optimization. The book is carefully written. Specific examples and concrete algorithms precede more abstract topics. Topics are clearly developed with a large number of numerical examples worked out in detail. Moreover, Linear Programming: Foundations and Extensions underscores the purpose of optimization: to solve practical problems on a computer. Accordingly, the book is coordinated with free efficient C programs that implement the major algorithms studied: -The two-phase simplex method; -The primal-dual simplex method; -The path-following interior-point method; -The homogeneous self-dual methods. In addition, there are online JAVA applets that illustrate various pivot rules and variants of the simplex m...
GPU acceleration of preconditioned solvers for ill-conditioned linear systems
Gupta, R.
2015-01-01
In this work we study the implementations of deflation and preconditioning techniques for solving ill-conditioned linear systems using iterative methods. Solving such systems can be a time-consuming process because of the jumps in the coefficients due to large difference in material properties. We
Ramírez-Bahena, Martha H; Vial, Ludovic; Lassalle, Florent; Diel, Benjamin; Chapulliot, David; Daubin, Vincent; Nesme, Xavier; Muller, Daniel
2014-04-01
Linear chromosomes are atypical in bacteria and likely a secondary trait derived from ancestral circular molecules. Within the Rhizobiaceae family, whose genome contains at least two chromosomes, a particularity of Agrobacterium fabrum (formerly A. tumefaciens) secondary chromosome (chromid) is to be linear and hairpin-ended thanks to the TelA protelomerase. Linear topology and telA distributions within this bacterial family was screened by pulse field gel electrophoresis and PCR. In A. rubi, A. larrymoorei, Rhizobium skierniewicense, A. viscosum, Agrobacterium sp. NCPPB 1650, and every genomospecies of the biovar 1/A. tumefaciens species complex (including R. pusense, A. radiobacter, A. fabrum, R. nepotum plus seven other unnamed genomospecies), linear chromid topologies were retrieved concomitantly with telA presence, whereas the remote species A. vitis, Allorhizobium undicola, Rhizobium rhizogenes and Ensifer meliloti harbored a circular chromid as well as no telA gene. Moreover, the telA phylogeny is congruent with that of recA used as a marker gene of the Agrobacterium phylogeny. Collectively, these findings strongly suggest that single acquisition of telA by an ancestor was the founding event of a large and diverse clade characterized by the presence of a linear chromid. This clade, characterized by unusual genome architecture, appears to be a relevant candidate to serve as a basis for a possible redefinition of the controversial Agrobacterium genus. In this respect, investigating telA in sequenced genomes allows to both ascertain the place of concerned strains into Agrobacterium spp. and their actual assignation to species/genomospecies in this genus. Copyright © 2014 Elsevier Inc. All rights reserved.
Solution of generalized shifted linear systems with complex symmetric matrices
International Nuclear Information System (INIS)
Sogabe, Tomohiro; Hoshi, Takeo; Zhang, Shao-Liang; Fujiwara, Takeo
2012-01-01
We develop the shifted COCG method [R. Takayama, T. Hoshi, T. Sogabe, S.-L. Zhang, T. Fujiwara, Linear algebraic calculation of Green’s function for large-scale electronic structure theory, Phys. Rev. B 73 (165108) (2006) 1–9] and the shifted WQMR method [T. Sogabe, T. Hoshi, S.-L. Zhang, T. Fujiwara, On a weighted quasi-residual minimization strategy of the QMR method for solving complex symmetric shifted linear systems, Electron. Trans. Numer. Anal. 31 (2008) 126–140] for solving generalized shifted linear systems with complex symmetric matrices that arise from the electronic structure theory. The complex symmetric Lanczos process with a suitable bilinear form plays an important role in the development of the methods. The numerical examples indicate that the methods are highly attractive when the inner linear systems can efficiently be solved.
Directory of Open Access Journals (Sweden)
Ching-Sung Wang
2016-09-01
Full Text Available Pitch Control plays a significant role for a large wind turbine. This study investigates a novel robust hydraulic pitch control system of a large wind turbine. The novel hydraulic pitch control system is driven by a novel high efficiency and high response hydraulic servo system. The pitch controller, designed by two degree-of-freedom (2-DOF motion control with feedback linearization, is developed to enhance the controllability and stability of the pitch control system. Furthermore, the full-scale testbed of the hydraulic pitch control system of a large wind turbine is developed for practically experimental verification. Besides, the wind turbine simulation software FAST is used to analyze the motion of the blade which results are given to the testbed as the disturbance load command. The 2-DOF pitch controller contains a feedforward controller with feedback linearization theory to overcome the nonlinearities of the system and a feedback controller to improve the system robustness for achieving the disturbance rejection. Consequently, the novel hydraulic pitch control system shows excellent path tracking performance in the experiments. Moreover, the robustness test with a simulated disturbance load generated by FAST is performed to validate the reliability of the proposed pitch control system.
Energy Technology Data Exchange (ETDEWEB)
Quirós Segovia, M.; Condés Ruiz, S.; Drápela, K.
2016-07-01
Aim of the study: The main objective of this study was to test Geographically Weighted Regression (GWR) for developing height-diameter curves for forests on a large scale and to compare it with Linear Mixed Models (LMM). Area of study: Monospecific stands of Pinus halepensis Mill. located in the region of Murcia (Southeast Spain). Materials and Methods: The dataset consisted of 230 sample plots (2582 trees) from the Third Spanish National Forest Inventory (SNFI) randomly split into training data (152 plots) and validation data (78 plots). Two different methodologies were used for modelling local (Petterson) and generalized height-diameter relationships (Cañadas I): GWR, with different bandwidths, and linear mixed models. Finally, the quality of the estimated models was compared throughout statistical analysis. Main results: In general, both LMM and GWR provide better prediction capability when applied to a generalized height-diameter function than when applied to a local one, with R2 values increasing from around 0.6 to 0.7 in the model validation. Bias and RMSE were also lower for the generalized function. However, error analysis showed that there were no large differences between these two methodologies, evidencing that GWR provides results which are as good as the more frequently used LMM methodology, at least when no additional measurements are available for calibrating. Research highlights: GWR is a type of spatial analysis for exploring spatially heterogeneous processes. GWR can model spatial variation in tree height-diameter relationship and its regression quality is comparable to LMM. The advantage of GWR over LMM is the possibility to determine the spatial location of every parameter without additional measurements. Abbreviations: GWR (Geographically Weighted Regression); LMM (Linear Mixed Model); SNFI (Spanish National Forest Inventory). (Author)
Olive, David J
2017-01-01
This text covers both multiple linear regression and some experimental design models. The text uses the response plot to visualize the model and to detect outliers, does not assume that the error distribution has a known parametric distribution, develops prediction intervals that work when the error distribution is unknown, suggests bootstrap hypothesis tests that may be useful for inference after variable selection, and develops prediction regions and large sample theory for the multivariate linear regression model that has m response variables. A relationship between multivariate prediction regions and confidence regions provides a simple way to bootstrap confidence regions. These confidence regions often provide a practical method for testing hypotheses. There is also a chapter on generalized linear models and generalized additive models. There are many R functions to produce response and residual plots, to simulate prediction intervals and hypothesis tests, to detect outliers, and to choose response trans...
Optimization theory for large systems
Lasdon, Leon S
2002-01-01
Important text examines most significant algorithms for optimizing large systems and clarifying relations between optimization procedures. Much data appear as charts and graphs and will be highly valuable to readers in selecting a method and estimating computer time and cost in problem-solving. Initial chapter on linear and nonlinear programming presents all necessary background for subjects covered in rest of book. Second chapter illustrates how large-scale mathematical programs arise from real-world problems. Appendixes. List of Symbols.
Navon, I. M.; Yu, Jian
A FORTRAN computer program is presented and documented applying the Turkel-Zwas explicit large time-step scheme to a hemispheric barotropic model with constraint restoration of integral invariants of the shallow-water equations. We then proceed to detail the algorithms embodied in the code EXSHALL in this paper, particularly algorithms related to the efficiency and stability of T-Z scheme and the quadratic constraint restoration method which is based on a variational approach. In particular we provide details about the high-latitude filtering, Shapiro filtering, and Robert filtering algorithms used in the code. We explain in detail the various subroutines in the EXSHALL code with emphasis on algorithms implemented in the code and present the flowcharts of some major subroutines. Finally, we provide a visual example illustrating a 4-day run using real initial data, along with a sample printout and graphic isoline contours of the height field and velocity fields.
Energy Technology Data Exchange (ETDEWEB)
Kozlowski, K.K. [Deutsches Elektronen-Synchrotron (DESY), Hamburg (Germany); Terras, V. [CNRS, ENS Lyon (France). Lab. de Physique
2010-12-15
We present a new method allowing us to derive the long-time and large-distance asymptotic behavior of the correlations functions of quantum integrable models from their exact representations. Starting from the form factor expansion of the correlation functions in finite volume, we explain how to reduce the complexity of the computation in the so-called interacting integrable models to the one appearing in free fermion equivalent models. We apply our method to the time-dependent zero-temperature current-current correlation function in the non-linear Schroedinger model and compute the first few terms in its asymptotic expansion. Our result goes beyond the conformal field theory based predictions: in the time-dependent case, other types of excitations than the ones on the Fermi surface contribute to the leading orders of the asymptotics. (orig.)
Murata, M; Uchida, T; Yang, Y; Lezhava, A; Kinashi, H
2011-04-01
We have comprehensively analyzed the linear chromosomes of Streptomyces griseus mutants constructed and kept in our laboratory. During this study, macrorestriction analysis of AseI and DraI fragments of mutant 402-2 suggested a large chromosomal inversion. The junctions of chromosomal inversion were cloned and sequenced and compared with the corresponding target sequences in the parent strain 2247. Consequently, a transposon-involved mechanism was revealed. Namely, a transposon originally located at the left target site was replicatively transposed to the right target site in an inverted direction, which generated a second copy and at the same time caused a 2.5-Mb chromosomal inversion. The involved transposon named TnSGR was grouped into a new subfamily of the resolvase-encoding Tn3 family transposons based on its gene organization. At the end, terminal diversity of S. griseus chromosomes is discussed by comparing the sequences of strains 2247 and IFO13350.
International Nuclear Information System (INIS)
Kozlowski, K.K.; Terras, V.
2010-12-01
We present a new method allowing us to derive the long-time and large-distance asymptotic behavior of the correlations functions of quantum integrable models from their exact representations. Starting from the form factor expansion of the correlation functions in finite volume, we explain how to reduce the complexity of the computation in the so-called interacting integrable models to the one appearing in free fermion equivalent models. We apply our method to the time-dependent zero-temperature current-current correlation function in the non-linear Schroedinger model and compute the first few terms in its asymptotic expansion. Our result goes beyond the conformal field theory based predictions: in the time-dependent case, other types of excitations than the ones on the Fermi surface contribute to the leading orders of the asymptotics. (orig.)
Energy Technology Data Exchange (ETDEWEB)
Kozlowski, K.K.
2010-12-15
Starting from the form factor expansion in finite volume, we derive the multidimensional generalization of the so-called Natte series for the zero-temperature, time and distance dependent reduced density matrix in the non-linear Schroedinger model. This representation allows one to read-off straightforwardly the long-time/large-distance asymptotic behavior of this correlator. Our method of analysis reduces the complexity of the computation of the asymptotic behavior of correlation functions in the so-called interacting integrable models, to the one appearing in free fermion equivalent models. We compute explicitly the first few terms appearing in the asymptotic expansion. Part of these terms stems from excitations lying away from the Fermi boundary, and hence go beyond what can be obtained by using the CFT/Luttinger liquid based predictions. (orig.)
Some Applications of Algebraic System Solving
Roanes-Lozano, Eugenio
2011-01-01
Technology and, in particular, computer algebra systems, allows us to change both the way we teach mathematics and the mathematical curriculum. Curiously enough, unlike what happens with linear system solving, algebraic system solving is not widely known. The aim of this paper is to show that, although the theory lying behind the "exact…
Solving applied mathematical problems with Matlab
Xue, Dingyu
2008-01-01
Computer Mathematics Language-An Overview. Fundamentals of MATLAB Programming. Calculus Problems. MATLAB Computations of Linear Algebra Problems. Integral Transforms and Complex Variable Functions. Solutions to Nonlinear Equations and Optimization Problems. MATLAB Solutions to Differential Equation Problems. Solving Interpolations and Approximations Problems. Solving Probability and Mathematical Statistics Problems. Nontraditional Solution Methods for Mathematical Problems.
Energy Technology Data Exchange (ETDEWEB)
Munehiro, H
1980-05-29
When driving the carriage of a printer through a rotating motor, there are problems regarding the limited accuracy of the carriage position due to rotation or contraction and ageing of the cable. In order to solve the problem, a direct drive system was proposed, in which the printer carriage is driven by a linear motor. If one wants to keep the motor circuit of such a motor compact, then the magnetic flux density in the air gap must be reduced or the motor travel must be reduced. It is the purpose of this invention to create an electrodynamic linear motor, which on the one hand is compact and light and on the other hand has a relatively high constant force over a large travel. The invention is characterised by the fact that magnetic fields of alternating polarity are generated at equal intervals in the magnetic field, and that the coil arrangement has 2 adjacent coils, whose size corresponds to half the length of each magnetic pole. A logic circuit is provided to select one of the two coils and to determine the direction of the current depending on the signals of a magnetic field sensor on the coil arrangement.
Topics in computational linear optimization
DEFF Research Database (Denmark)
Hultberg, Tim Helge
2000-01-01
Linear optimization has been an active area of research ever since the pioneering work of G. Dantzig more than 50 years ago. This research has produced a long sequence of practical as well as theoretical improvements of the solution techniques avilable for solving linear optimization problems...... of high quality solvers and the use of algebraic modelling systems to handle the communication between the modeller and the solver. This dissertation features four topics in computational linear optimization: A) automatic reformulation of mixed 0/1 linear programs, B) direct solution of sparse unsymmetric...... systems of linear equations, C) reduction of linear programs and D) integration of algebraic modelling of linear optimization problems in C++. Each of these topics is treated in a separate paper included in this dissertation. The efficiency of solving mixed 0-1 linear programs by linear programming based...
Energy Technology Data Exchange (ETDEWEB)
Ahangarianabhari, Mahdi; Macera, Daniele [Politecnico di Milano, Department of Electronics Engineering, Information Science and Bioengineering, P.za L. da Vinci 32, 20133 Milano (Italy); National Institute of Nuclear Physics, INFN sez. Milano (Italy); Bertuccio, Giuseppe, E-mail: Giuseppe.Bertuccio@polimi.it [Politecnico di Milano, Department of Electronics Engineering, Information Science and Bioengineering, P.za L. da Vinci 32, 20133 Milano (Italy); National Institute of Nuclear Physics, INFN sez. Milano (Italy); Malcovati, Piero; Grassi, Marco [University of Pavia, Department of Electrical Engineering, and National Institute of Nuclear Physics, INFN sez. Pavia, Pavia (Italy)
2015-01-11
We present the design and the first experimental characterization of VEGA, an Application Specific Integrated Circuit (ASIC) designed to read out large area monolithic linear Silicon Drift Detectors (SDD’s). VEGA consists of an analog and a digital/mixed-signal section to accomplish all the functionalities and specifications required for high resolution X-ray spectroscopy in the energy range between 500 eV and 50 keV. The analog section includes a charge sensitive preamplifier, a shaper with 3-bit digitally selectable shaping times from 1.6 µs to 6.6 µs and a peak stretcher/sample-and-hold stage. The digital/mixed-signal section includes an amplitude discriminator with coarse and fine threshold level setting, a peak discriminator and a logic circuit to fulfill pile-up rejection, signal sampling, trigger generation, channel reset and the preamplifier and discriminators disabling functionalities. A Serial Peripherical Interface (SPI) is integrated in VEGA for loading and storing all configuration parameters in an internal register within few microseconds. The VEGA ASIC has been designed and manufactured in 0.35 µm CMOS mixed-signal technology in single and 32 channel versions with dimensions of 200 µm×500 µm per channel. A minimum intrinsic Equivalent Noise Charge (ENC) of 12 electrons r.m.s. at 3.6 µs peaking time and room temperature is measured and the linearity error is between −0.9% and +0.6% in the whole input energy range. The total power consumption is 481 µW and 420 µW per channel for the single and 32 channels version, respectively. A comparison with other ASICs for X-ray SDD’s shows that VEGA has a suitable low noise and offers high functionality as ADC-ready signal processing but at a power consumption that is a factor of four lower than other similar existing ASICs.
Review on solving the forward problem in EEG source analysis
Directory of Open Access Journals (Sweden)
Vergult Anneleen
2007-11-01
Full Text Available Abstract Background The aim of electroencephalogram (EEG source localization is to find the brain areas responsible for EEG waves of interest. It consists of solving forward and inverse problems. The forward problem is solved by starting from a given electrical source and calculating the potentials at the electrodes. These evaluations are necessary to solve the inverse problem which is defined as finding brain sources which are responsible for the measured potentials at the EEG electrodes. Methods While other reviews give an extensive summary of the both forward and inverse problem, this review article focuses on different aspects of solving the forward problem and it is intended for newcomers in this research field. Results It starts with focusing on the generators of the EEG: the post-synaptic potentials in the apical dendrites of pyramidal neurons. These cells generate an extracellular current which can be modeled by Poisson's differential equation, and Neumann and Dirichlet boundary conditions. The compartments in which these currents flow can be anisotropic (e.g. skull and white matter. In a three-shell spherical head model an analytical expression exists to solve the forward problem. During the last two decades researchers have tried to solve Poisson's equation in a realistically shaped head model obtained from 3D medical images, which requires numerical methods. The following methods are compared with each other: the boundary element method (BEM, the finite element method (FEM and the finite difference method (FDM. In the last two methods anisotropic conducting compartments can conveniently be introduced. Then the focus will be set on the use of reciprocity in EEG source localization. It is introduced to speed up the forward calculations which are here performed for each electrode position rather than for each dipole position. Solving Poisson's equation utilizing FEM and FDM corresponds to solving a large sparse linear system. Iterative
Superconducting linear accelerator cryostat
International Nuclear Information System (INIS)
Ben-Zvi, I.; Elkonin, B.V.; Sokolowski, J.S.
1984-01-01
A large vertical cryostat for a superconducting linear accelerator using quarter wave resonators has been developed. The essential technical details, operational experience and performance are described. (author)
Energy Technology Data Exchange (ETDEWEB)
Xie, Jiazhuo; Zhang, Kun; Zhao, Qinghua [College of Chemistry and Material Science, Shandong Agricultural University, 61 Daizong Street, Tai' an 271018 (China); Wang, Qingguo, E-mail: wqgyyy@126.com [College of Food Science and Engineering, Shandong Agricultural University, 61 Daizong Street, Tai' an 271018 (China); Xu, Jing, E-mail: jiaxu@sdau.edu.cn [College of Chemistry and Material Science, Shandong Agricultural University, 61 Daizong Street, Tai' an 271018 (China)
2016-11-15
Novel LDH intercalated with organic aliphatic long-chain anion was large-scale synthesized innovatively by high-energy ball milling in one pot. The linear low density polyethylene (LLDPE)/layered double hydroxides (LDH) composite films with enhanced heat retention, thermal, mechanical, optical and water vapor barrier properties were fabricated by melt blending and blowing process. FT IR, XRD, SEM results show that LDH particles were dispersed uniformly in the LLDPE composite films. Particularly, LLDPE composite film with 1% LDH exhibited the optimal performance among all the composite films with a 60.36% enhancement in the water vapor barrier property and a 45.73 °C increase in the temperature of maximum mass loss rate compared with pure LLDPE film. Furthermore, the improved infrared absorbance (1180–914 cm{sup −1}) of LLDPE/LDH films revealed the significant enhancement of heat retention. Therefore, this study prompts the application of LLDPE/LDH films as agricultural films with superior heat retention. - Graphical abstract: The fabrication process of LLDPE/LDH composite films. - Highlights: • LDH with basal spacing of 4.07 nm was synthesized by high-energy ball milling. • LLDPE composite films with homogeneous LDH dispersion were fabricated. • The properties of LLDPE/LDH composite films were improved. • LLDPE/LDH composite films show superior heat retention property.
Energy Technology Data Exchange (ETDEWEB)
Mapas, Jose Kenneth D.; Thomay, Tim; Cartwright, Alexander N.; Ilavsky, Jan; Rzayev, Javid
2016-05-05
Block copolymer (BCP) derived periodic nanostructures with domain sizes larger than 150 nm present a versatile platform for the fabrication of photonic materials. So far, the access to such materials has been limited to highly synthetically involved protocols. Herein, we report a simple, “user-friendly” method for the preparation of ultrahigh molecular weight linear poly(solketal methacrylate-b-styrene) block copolymers by a combination of Cu-wire-mediated ATRP and RAFT polymerizations. The synthesized copolymers with molecular weights up to 1.6 million g/mol and moderate dispersities readily assemble into highly ordered cylindrical or lamella microstructures with domain sizes as large as 292 nm, as determined by ultra-small-angle x-ray scattering and scanning electron microscopy analyses. Solvent cast films of the synthesized block copolymers exhibit stop bands in the visible spectrum correlated to their domain spacings. The described method opens new avenues for facilitated fabrication and the advancement of fundamental understanding of BCP-derived photonic nanomaterials for a variety of applications.
International Nuclear Information System (INIS)
Xie, Jiazhuo; Zhang, Kun; Zhao, Qinghua; Wang, Qingguo; Xu, Jing
2016-01-01
Novel LDH intercalated with organic aliphatic long-chain anion was large-scale synthesized innovatively by high-energy ball milling in one pot. The linear low density polyethylene (LLDPE)/layered double hydroxides (LDH) composite films with enhanced heat retention, thermal, mechanical, optical and water vapor barrier properties were fabricated by melt blending and blowing process. FT IR, XRD, SEM results show that LDH particles were dispersed uniformly in the LLDPE composite films. Particularly, LLDPE composite film with 1% LDH exhibited the optimal performance among all the composite films with a 60.36% enhancement in the water vapor barrier property and a 45.73 °C increase in the temperature of maximum mass loss rate compared with pure LLDPE film. Furthermore, the improved infrared absorbance (1180–914 cm −1 ) of LLDPE/LDH films revealed the significant enhancement of heat retention. Therefore, this study prompts the application of LLDPE/LDH films as agricultural films with superior heat retention. - Graphical abstract: The fabrication process of LLDPE/LDH composite films. - Highlights: • LDH with basal spacing of 4.07 nm was synthesized by high-energy ball milling. • LLDPE composite films with homogeneous LDH dispersion were fabricated. • The properties of LLDPE/LDH composite films were improved. • LLDPE/LDH composite films show superior heat retention property.
Patterson, Brian M; Havrilla, George J
2006-11-01
The number of techniques and instruments available for Fourier transform infrared (FT-IR) microspectroscopic imaging has grown significantly over the past few years. Attenuated total internal reflectance (ATR) FT-IR microspectroscopy reduces sample preparation time and has simplified the analysis of many difficult samples. FT-IR imaging has become a powerful analytical tool using either a focal plane array or a linear array detector, especially when coupled with a chemometric analysis package. The field of view of the ATR-IR microspectroscopic imaging area can be greatly increased from 300 x 300 microm to 2500 x 2500 microm using a larger internal reflection element of 12.5 mm radius instead of the typical 1.5 mm radius. This gives an area increase of 70x before aberrant effects become too great. Parameters evaluated include the change in penetration depth as a function of beam displacement, measurements of the active area, magnification factor, and change in spatial resolution over the imaging area. Drawbacks such as large file size will also be discussed. This technique has been successfully applied to the FT-IR imaging of polydimethylsiloxane foam cross-sections, latent human fingerprints, and a model inorganic mixture, which demonstrates the usefulness of the method for pharmaceuticals.
Energy Technology Data Exchange (ETDEWEB)
Wiedemann, H.
1981-11-01
Since no linear colliders have been built yet it is difficult to know at what energy the linear cost scaling of linear colliders drops below the quadratic scaling of storage rings. There is, however, no doubt that a linear collider facility for a center of mass energy above say 500 GeV is significantly cheaper than an equivalent storage ring. In order to make the linear collider principle feasible at very high energies a number of problems have to be solved. There are two kinds of problems: one which is related to the feasibility of the principle and the other kind of problems is associated with minimizing the cost of constructing and operating such a facility. This lecture series describes the problems and possible solutions. Since the real test of a principle requires the construction of a prototype I will in the last chapter describe the SLC project at the Stanford Linear Accelerator Center.
International Nuclear Information System (INIS)
Wiedemann, H.
1981-11-01
Since no linear colliders have been built yet it is difficult to know at what energy the linear cost scaling of linear colliders drops below the quadratic scaling of storage rings. There is, however, no doubt that a linear collider facility for a center of mass energy above say 500 GeV is significantly cheaper than an equivalent storage ring. In order to make the linear collider principle feasible at very high energies a number of problems have to be solved. There are two kinds of problems: one which is related to the feasibility of the principle and the other kind of problems is associated with minimizing the cost of constructing and operating such a facility. This lecture series describes the problems and possible solutions. Since the real test of a principle requires the construction of a prototype I will in the last chapter describe the SLC project at the Stanford Linear Accelerator Center
Estimation and variable selection for generalized additive partial linear models
Wang, Li
2011-08-01
We study generalized additive partial linear models, proposing the use of polynomial spline smoothing for estimation of nonparametric functions, and deriving quasi-likelihood based estimators for the linear parameters. We establish asymptotic normality for the estimators of the parametric components. The procedure avoids solving large systems of equations as in kernel-based procedures and thus results in gains in computational simplicity. We further develop a class of variable selection procedures for the linear parameters by employing a nonconcave penalized quasi-likelihood, which is shown to have an asymptotic oracle property. Monte Carlo simulations and an empirical example are presented for illustration. © Institute of Mathematical Statistics, 2011.
Linear local stability of electrostatic drift modes in helical systems
International Nuclear Information System (INIS)
Yamagishi, O.; Nakajima, N.; Sugama, H.; Nakamura, Y.
2003-01-01
We investigate the stability of the drift wave in helical systems. For this purpose, we solve the linear local gyrokinetic-Poisson equation, in the electrostatic regime. As a model of helical plasmas, Large helical Device (LHD) is considered. The equation we apply is rather exact in the framework of linear gyrokinetic theory, where only the approximation is the ballooning representation. In this paper, we consider only collisionless cases. All the frequency regime can be naturally reated without any assumptions, and in such cases, ion temperature gradient modes (ITG), trapped electron modes (TEM), and electron temperature gradient modes (ETG) are expected to become unstable linearly independently. (orig.)
Masuda, Y; Misztal, I; Legarra, A; Tsuruta, S; Lourenco, D A L; Fragomeni, B O; Aguilar, I
2017-01-01
This paper evaluates an efficient implementation to multiply the inverse of a numerator relationship matrix for genotyped animals () by a vector (). The computation is required for solving mixed model equations in single-step genomic BLUP (ssGBLUP) with the preconditioned conjugate gradient (PCG). The inverse can be decomposed into sparse matrices that are blocks of the sparse inverse of a numerator relationship matrix () including genotyped animals and their ancestors. The elements of were rapidly calculated with the Henderson's rule and stored as sparse matrices in memory. Implementation of was by a series of sparse matrix-vector multiplications. Diagonal elements of , which were required as preconditioners in PCG, were approximated with a Monte Carlo method using 1,000 samples. The efficient implementation of was compared with explicit inversion of with 3 data sets including about 15,000, 81,000, and 570,000 genotyped animals selected from populations with 213,000, 8.2 million, and 10.7 million pedigree animals, respectively. The explicit inversion required 1.8 GB, 49 GB, and 2,415 GB (estimated) of memory, respectively, and 42 s, 56 min, and 13.5 d (estimated), respectively, for the computations. The efficient implementation required <1 MB, 2.9 GB, and 2.3 GB of memory, respectively, and <1 sec, 3 min, and 5 min, respectively, for setting up. Only <1 sec was required for the multiplication in each PCG iteration for any data sets. When the equations in ssGBLUP are solved with the PCG algorithm, is no longer a limiting factor in the computations.
Schüle, Steffen Andreas; Gabriel, Katharina M A; Bolte, Gabriele
2017-06-01
The environmental justice framework states that besides environmental burdens also resources may be social unequally distributed both on the individual and on the neighbourhood level. This ecological study investigated whether neighbourhood socioeconomic position (SEP) was associated with neighbourhood public green space availability in a large German city with more than 1 million inhabitants. Two different measures were defined for green space availability. Firstly, percentage of green space within neighbourhoods was calculated with the additional consideration of various buffers around the boundaries. Secondly, percentage of green space was calculated based on various radii around the neighbourhood centroid. An index of neighbourhood SEP was calculated with principal component analysis. Log-gamma regression from the group of generalized linear models was applied in order to consider the non-normal distribution of the response variable. All models were adjusted for population density. Low neighbourhood SEP was associated with decreasing neighbourhood green space availability including 200m up to 1000m buffers around the neighbourhood boundaries. Low neighbourhood SEP was also associated with decreasing green space availability based on catchment areas measured from neighbourhood centroids with different radii (1000m up to 3000 m). With an increasing radius the strength of the associations decreased. Social unequally distributed green space may amplify environmental health inequalities in an urban context. Thus, the identification of vulnerable neighbourhoods and population groups plays an important role for epidemiological research and healthy city planning. As a methodical aspect, log-gamma regression offers an adequate parametric modelling strategy for positively distributed environmental variables. Copyright © 2017 Elsevier GmbH. All rights reserved.
Skinner, Mark F; Hopwood, David
2004-03-01
Repetitive linear enamel hypoplasia (rLEH) is often observed in recent large-bodied apes from Africa and Asia as well as Mid- to Late Miocene sites from Spain to China. The ubiquity and periodicity of rLEH are not understood. Its potential as an ontogenetic marker of developmental stress in threatened species (as well as their ancient relatives) makes rLEH an important if enigmatic problem. We report research designed to show the periodicity of rLEH among West African Pan troglodytes (12 male, 32 female), Gorilla gorilla (10 male, 10 female), and Bornean and Sumatran Pongo pygmaeus (11 male, 9 female, 9 unknown) from collections in Europe. Two methods were employed. In the common chimpanzees and gorillas, the space between adjacent, macroscopically visible LEH grooves on teeth with two or more episodes was expressed as an absolute measure and as a ratio of complete unworn crown height. In the orangutans, the number of perikymata between episode onsets, as well as duration of rLEH, was determined from scanning electron micrographs of casts of incisors and canines. We conclude that stress in the form of LEH commences as early as 2.5 years of age in all taxa and lasts for several years, and even longer in orangutans; the stress is not chronic but episodic; the stressor has a strong tendency to occur in pulses of two occurrences each; and large apes from both land masses exhibit rLEH with an average periodicity of 6 months (or multiples thereof; Sumatran orangutans seem to show only annual stress), but this needs further research. This is supported by evidence of spacing between rLEH as well as perikymata counts. Duration of stress in orangutans averages about 6 weeks. Finally, the semiannual stressor transcends geographic and temporal boundaries, and is attributed to regular moisture cycles associated with the intertropical convergence zone modified by the monsoon. While seasonal cycles can influence both disease and nutritional stress, it is likely the combination of
International Nuclear Information System (INIS)
Bracalente, Candelaria; Ibañez, Irene L.; Molinari, Beatriz; Palmieri, Mónica; Kreiner, Andrés; Valda, Alejandro
2013-01-01
Purpose: To evaluate the cell response to DNA double-strand breaks induced by low and high linear energy transfer (LET) radiations when the catalytic subunit of DNA-dependent protein kinase (DNA-PKcs), an essential protein of the nonhomologous end-joining repair pathway, lacks kinase activity. Methods and Materials: CHO10B2, a Chinese hamster ovary cell line, and its derived radiosensitive mutant cell line, irs-20, lacking DNA-PKcs activity, were evaluated after 0 to 3 Gy of γ-rays, plateau and Bragg peak protons, and lithium beams by clonogenic assay, and as a measurement of double-strand breaks, phosphorylated H2AX (γH2AX) foci number and size were quantified by immunocytofluorescence. Results: Irs-20 exhibited greater radiosensitivity and a higher amount of γH2AX foci than CHO10B2 at 6 hours after irradiation for all types of radiations. Remarkably, CHO10B2 and irs-20 maintained their difference in radiosensitivity after high-LET radiation. Six hours after low-LET radiations, irs-20 did not reach basal levels of γH2AX at high doses, whereas CHO10B2 recovered basal levels for all doses. After high-LET radiation, only CHO10B2 exhibited a reduction in γH2AX foci, but it never reached basal levels. Persistent foci in irs-20 confirmed a repair deficiency. Interestingly, after 30 minutes of high-LET radiation both cell lines exhibited large foci (size >0.9 μm 2 ) related to the damage nature, whereas at 6 hours irs-20 showed a higher amount of large foci than CHO10B2, with a 7-fold increase at 3 Gy, that could also be associated to radiosensitivity. Conclusions: We demonstrated, for the first time, an association between deficient DNA-PKcs activity and not only high levels of H2AX phosphorylation but also persistence and size increase of γH2AX foci after high-LET irradiation
Energy Technology Data Exchange (ETDEWEB)
Bracalente, Candelaria; Ibañez, Irene L. [Departamento de Micro y Nanotecnología, Comisión Nacional de Energía Atómica, San Martín, Buenos Aires (Argentina); Consejo Nacional de Investigaciones Científicas y Técnicas, Buenos Aires (Argentina); Molinari, Beatriz [Departamento de Radiobiología, Comisión Nacional de Energía Atómica, San Martín, Buenos Aires (Argentina); Consejo Nacional de Investigaciones Científicas y Técnicas, Buenos Aires (Argentina); Palmieri, Mónica [Facultad de Ciencias Exactas y Naturales, Universidad de Buenos Aires, Buenos Aires (Argentina); Kreiner, Andrés [Consejo Nacional de Investigaciones Científicas y Técnicas, Buenos Aires (Argentina); Gerencia de Investigación y Aplicaciones, Comisión Nacional de Energía Atómica, San Martín, Buenos Aires (Argentina); Escuela de Ciencia y Tecnología, Universidad Nacional de San Martín, San Martín, Buenos Aires (Argentina); Valda, Alejandro [Escuela de Ciencia y Tecnología, Universidad Nacional de San Martín, San Martín, Buenos Aires (Argentina); and others
2013-11-15
Purpose: To evaluate the cell response to DNA double-strand breaks induced by low and high linear energy transfer (LET) radiations when the catalytic subunit of DNA-dependent protein kinase (DNA-PKcs), an essential protein of the nonhomologous end-joining repair pathway, lacks kinase activity. Methods and Materials: CHO10B2, a Chinese hamster ovary cell line, and its derived radiosensitive mutant cell line, irs-20, lacking DNA-PKcs activity, were evaluated after 0 to 3 Gy of γ-rays, plateau and Bragg peak protons, and lithium beams by clonogenic assay, and as a measurement of double-strand breaks, phosphorylated H2AX (γH2AX) foci number and size were quantified by immunocytofluorescence. Results: Irs-20 exhibited greater radiosensitivity and a higher amount of γH2AX foci than CHO10B2 at 6 hours after irradiation for all types of radiations. Remarkably, CHO10B2 and irs-20 maintained their difference in radiosensitivity after high-LET radiation. Six hours after low-LET radiations, irs-20 did not reach basal levels of γH2AX at high doses, whereas CHO10B2 recovered basal levels for all doses. After high-LET radiation, only CHO10B2 exhibited a reduction in γH2AX foci, but it never reached basal levels. Persistent foci in irs-20 confirmed a repair deficiency. Interestingly, after 30 minutes of high-LET radiation both cell lines exhibited large foci (size >0.9 μm{sup 2}) related to the damage nature, whereas at 6 hours irs-20 showed a higher amount of large foci than CHO10B2, with a 7-fold increase at 3 Gy, that could also be associated to radiosensitivity. Conclusions: We demonstrated, for the first time, an association between deficient DNA-PKcs activity and not only high levels of H2AX phosphorylation but also persistence and size increase of γH2AX foci after high-LET irradiation.
Shilov, Georgi E
1977-01-01
Covers determinants, linear spaces, systems of linear equations, linear functions of a vector argument, coordinate transformations, the canonical form of the matrix of a linear operator, bilinear and quadratic forms, Euclidean spaces, unitary spaces, quadratic forms in Euclidean and unitary spaces, finite-dimensional space. Problems with hints and answers.
International Nuclear Information System (INIS)
Gunyasu, Kenzo; Hiramoto, Tsuneyuki; Tanimoto, Mitsumori; Osano, Minetada
2002-01-01
We describe a new method for solving large-scale system of linear equations resulting from discretization of ordinary differential equation and partial differential equation directly. This new method effectively reduces the memory capacity requirements and computing time problems for analyses using finite difference method and finite element method. In this paper we have tried to solve one-million linear equations directly for the case that initial displacement and boundary displacement are known about the finite difference scheme of second order inhomogeneous differential equation for vibration of a 10 story structure. Excellent results were got. (author)
ALPS - A LINEAR PROGRAM SOLVER
Viterna, L. A.
1994-01-01
Linear programming is a widely-used engineering and management tool. Scheduling, resource allocation, and production planning are all well-known applications of linear programs (LP's). Most LP's are too large to be solved by hand, so over the decades many computer codes for solving LP's have been developed. ALPS, A Linear Program Solver, is a full-featured LP analysis program. ALPS can solve plain linear programs as well as more complicated mixed integer and pure integer programs. ALPS also contains an efficient solution technique for pure binary (0-1 integer) programs. One of the many weaknesses of LP solvers is the lack of interaction with the user. ALPS is a menu-driven program with no special commands or keywords to learn. In addition, ALPS contains a full-screen editor to enter and maintain the LP formulation. These formulations can be written to and read from plain ASCII files for portability. For those less experienced in LP formulation, ALPS contains a problem "parser" which checks the formulation for errors. ALPS creates fully formatted, readable reports that can be sent to a printer or output file. ALPS is written entirely in IBM's APL2/PC product, Version 1.01. The APL2 workspace containing all the ALPS code can be run on any APL2/PC system (AT or 386). On a 32-bit system, this configuration can take advantage of all extended memory. The user can also examine and modify the ALPS code. The APL2 workspace has also been "packed" to be run on any DOS system (without APL2) as a stand-alone "EXE" file, but has limited memory capacity on a 640K system. A numeric coprocessor (80X87) is optional but recommended. The standard distribution medium for ALPS is a 5.25 inch 360K MS-DOS format diskette. IBM, IBM PC and IBM APL2 are registered trademarks of International Business Machines Corporation. MS-DOS is a registered trademark of Microsoft Corporation.
Linear versus non-linear supersymmetry, in general
Energy Technology Data Exchange (ETDEWEB)
Ferrara, Sergio [Theoretical Physics Department, CERN,CH-1211 Geneva 23 (Switzerland); INFN - Laboratori Nazionali di Frascati,Via Enrico Fermi 40, I-00044 Frascati (Italy); Department of Physics and Astronomy, UniversityC.L.A.,Los Angeles, CA 90095-1547 (United States); Kallosh, Renata [SITP and Department of Physics, Stanford University,Stanford, California 94305 (United States); Proeyen, Antoine Van [Institute for Theoretical Physics, Katholieke Universiteit Leuven,Celestijnenlaan 200D, B-3001 Leuven (Belgium); Wrase, Timm [Institute for Theoretical Physics, Technische Universität Wien,Wiedner Hauptstr. 8-10, A-1040 Vienna (Austria)
2016-04-12
We study superconformal and supergravity models with constrained superfields. The underlying version of such models with all unconstrained superfields and linearly realized supersymmetry is presented here, in addition to the physical multiplets there are Lagrange multiplier (LM) superfields. Once the equations of motion for the LM superfields are solved, some of the physical superfields become constrained. The linear supersymmetry of the original models becomes non-linearly realized, its exact form can be deduced from the original linear supersymmetry. Known examples of constrained superfields are shown to require the following LM’s: chiral superfields, linear superfields, general complex superfields, some of them are multiplets with a spin.
Linear versus non-linear supersymmetry, in general
International Nuclear Information System (INIS)
Ferrara, Sergio; Kallosh, Renata; Proeyen, Antoine Van; Wrase, Timm
2016-01-01
We study superconformal and supergravity models with constrained superfields. The underlying version of such models with all unconstrained superfields and linearly realized supersymmetry is presented here, in addition to the physical multiplets there are Lagrange multiplier (LM) superfields. Once the equations of motion for the LM superfields are solved, some of the physical superfields become constrained. The linear supersymmetry of the original models becomes non-linearly realized, its exact form can be deduced from the original linear supersymmetry. Known examples of constrained superfields are shown to require the following LM’s: chiral superfields, linear superfields, general complex superfields, some of them are multiplets with a spin.
Parallel Algorithm Solves Coupled Differential Equations
Hayashi, A.
1987-01-01
Numerical methods adapted to concurrent processing. Algorithm solves set of coupled partial differential equations by numerical integration. Adapted to run on hypercube computer, algorithm separates problem into smaller problems solved concurrently. Increase in computing speed with concurrent processing over that achievable with conventional sequential processing appreciable, especially for large problems.
A Novel Approach for Solving Semidefinite Programs
Directory of Open Access Journals (Sweden)
Hong-Wei Jiao
2014-01-01
Full Text Available A novel linearizing alternating direction augmented Lagrangian approach is proposed for effectively solving semidefinite programs (SDP. For every iteration, by fixing the other variables, the proposed approach alternatively optimizes the dual variables and the dual slack variables; then the primal variables, that is, Lagrange multipliers, are updated. In addition, the proposed approach renews all the variables in closed forms without solving any system of linear equations. Global convergence of the proposed approach is proved under mild conditions, and two numerical problems are given to demonstrate the effectiveness of the presented approach.
Carlson, H. W.
1979-01-01
A new linearized-theory pressure-coefficient formulation was studied. The new formulation is intended to provide more accurate estimates of detailed pressure loadings for improved stability analysis and for analysis of critical structural design conditions. The approach is based on the use of oblique-shock and Prandtl-Meyer expansion relationships for accurate representation of the variation of pressures with surface slopes in two-dimensional flow and linearized-theory perturbation velocities for evaluation of local three-dimensional aerodynamic interference effects. The applicability and limitations of the modification to linearized theory are illustrated through comparisons with experimental pressure distributions for delta wings covering a Mach number range from 1.45 to 4.60 and angles of attack from 0 to 25 degrees.
Linear Programming and Network Flows
Bazaraa, Mokhtar S; Sherali, Hanif D
2011-01-01
The authoritative guide to modeling and solving complex problems with linear programming-extensively revised, expanded, and updated The only book to treat both linear programming techniques and network flows under one cover, Linear Programming and Network Flows, Fourth Edition has been completely updated with the latest developments on the topic. This new edition continues to successfully emphasize modeling concepts, the design and analysis of algorithms, and implementation strategies for problems in a variety of fields, including industrial engineering, management science, operations research
Solving Conic Systems via Projection and Rescaling
Pena, Javier; Soheili, Negar
2015-01-01
We propose a simple projection and rescaling algorithm to solve the feasibility problem \\[ \\text{ find } x \\in L \\cap \\Omega, \\] where $L$ and $\\Omega$ are respectively a linear subspace and the interior of a symmetric cone in a finite-dimensional vector space $V$. This projection and rescaling algorithm is inspired by previous work on rescaled versions of the perceptron algorithm and by Chubanov's projection-based method for linear feasibility problems. As in these predecessors, each main it...
Fundamentals of linear algebra
Dash, Rajani Ballav
2008-01-01
FUNDAMENTALS OF LINEAR ALGEBRA is a comprehensive Text Book, which can be used by students and teachers of All Indian Universities. The Text has easy, understandable form and covers all topics of UGC Curriculum. There are lots of worked out examples which helps the students in solving the problems without anybody's help. The Problem sets have been designed keeping in view of the questions asked in different examinations.
International Nuclear Information System (INIS)
Scheffel, J.
1984-03-01
The linear Grad-Shafranov equation for a toroidal, axisymmetric plasma is solved analytically. Exact solutions are given in terms of confluent hyper-geometric functions. As an alternative, simple and accurate WKBJ solutions are presented. With parabolic pressure profiles, both hollow and peaked toroidal current density profiles are obtained. As an example the equilibrium of a z-pinch with a square-shaped cross section is derived.(author)
Li, Jiaru; Joubert-Doriol, Loïc; Izmaylov, Artur F.
2017-08-01
We investigate geometric phase (GP) effects in nonadiabatic transitions through a conical intersection (CI) in an N-dimensional linear vibronic coupling (ND-LVC) model. This model allows for the coordinate transformation encompassing all nonadiabatic effects within a two-dimensional (2D) subsystem, while the other N - 2 dimensions form a system of uncoupled harmonic oscillators identical for both electronic states and coupled bi-linearly with the subsystem coordinates. The 2D subsystem governs ultra-fast nonadiabatic dynamics through the CI and provides a convenient model for studying GP effects. Parameters of the original ND-LVC model define the Hamiltonian of the transformed 2D subsystem and thus influence GP effects directly. Our analysis reveals what values of ND-LVC parameters can introduce symmetry breaking in the 2D subsystem that diminishes GP effects.
Szyld, D. B.
1984-01-01
A brief description of the Model of the World Economy implemented at the Institute for Economic Analysis is presented, together with our experience in converting the software to vector code. For each time period, the model is reduced to a linear system of over 2000 variables. The matrix of coefficients has a bordered block diagonal structure, and we show how some of the matrix operations can be carried out on all diagonal blocks at once.
Singh, Rahul; Shukla, K. K.; Kumar, A.; Okram, G. S.; Singh, D.; Ganeshan, V.; Lakhani, Archana; Ghosh, A. K.; Chatterjee, Sandip
2016-09-01
Magnetoresistance (MR), thermo power, magnetization and Hall effect measurements have been performed on Co-doped Bi2Se3 topological insulators. The undoped sample shows that the maximum MR as a destructive interference due to a π-Berry phase leads to a decrease of MR. As the Co is doped, the linearity in MR is increased. The observed MR of Bi2Se3 can be explained with the classical model. The low temperature MR behavior of Co doped samples cannot be explained with the same model, but can be explained with the quantum linear MR model. Magnetization behavior indicates the establishment of ferromagnetic ordering with Co doping. Hall effect data also supports the establishment of ferromagnetic ordering in Co-doped Bi2Se3 samples by showing the anomalous Hall effect. Furthermore, when spectral weight suppression is insignificant, Bi2Se3 behaves as a dilute magnetic semiconductor. Moreover, the maximum power factor is observed when time reversal symmetry (TRS) is maintained. As the TRS is broken the power factor value is decreased, which indicates that with the rise of Dirac cone above the Fermi level the anomalous Hall effect and linearity in MR increase and the power factor decreases.
Application of Nearly Linear Solvers to Electric Power System Computation
Grant, Lisa L.
To meet the future needs of the electric power system, improvements need to be made in the areas of power system algorithms, simulation, and modeling, specifically to achieve a time frame that is useful to industry. If power system time-domain simulations could run in real-time, then system operators would have situational awareness to implement online control and avoid cascading failures, significantly improving power system reliability. Several power system applications rely on the solution of a very large linear system. As the demands on power systems continue to grow, there is a greater computational complexity involved in solving these large linear systems within reasonable time. This project expands on the current work in fast linear solvers, developed for solving symmetric and diagonally dominant linear systems, in order to produce power system specific methods that can be solved in nearly-linear run times. The work explores a new theoretical method that is based on ideas in graph theory and combinatorics. The technique builds a chain of progressively smaller approximate systems with preconditioners based on the system's low stretch spanning tree. The method is compared to traditional linear solvers and shown to reduce the time and iterations required for an accurate solution, especially as the system size increases. A simulation validation is performed, comparing the solution capabilities of the chain method to LU factorization, which is the standard linear solver for power flow. The chain method was successfully demonstrated to produce accurate solutions for power flow simulation on a number of IEEE test cases, and a discussion on how to further improve the method's speed and accuracy is included.
International Nuclear Information System (INIS)
Richter, B.
1985-01-01
A report is given on the goals and progress of the SLAC Linear Collider. The author discusses the status of the machine and the detectors and give an overview of the physics which can be done at this new facility. He also gives some ideas on how (and why) large linear colliders of the future should be built
A method for solving neutron transport equation
International Nuclear Information System (INIS)
Dimitrijevic, Z.
1993-01-01
The procedure for solving the transport equation by directly integrating for case one-dimensional uniform multigroup medium is shown. The solution is expressed in terms of linear combination of function H n (x,μ), and the coefficient is determined from given conditions. The solution is applied for homogeneous slab of critical thickness. (author)
Relaxation Methods for Strictly Convex Regularizations of Piecewise Linear Programs
International Nuclear Information System (INIS)
Kiwiel, K. C.
1998-01-01
We give an algorithm for minimizing the sum of a strictly convex function and a convex piecewise linear function. It extends several dual coordinate ascent methods for large-scale linearly constrained problems that occur in entropy maximization, quadratic programming, and network flows. In particular, it may solve exact penalty versions of such (possibly inconsistent) problems, and subproblems of bundle methods for nondifferentiable optimization. It is simple, can exploit sparsity, and in certain cases is highly parallelizable. Its global convergence is established in the recent framework of B -functions (generalized Bregman functions)
International Nuclear Information System (INIS)
Suwono.
1978-01-01
A linear gate providing a variable gate duration from 0,40μsec to 4μsec was developed. The electronic circuity consists of a linear circuit and an enable circuit. The input signal can be either unipolar or bipolar. If the input signal is bipolar, the negative portion will be filtered. The operation of the linear gate is controlled by the application of a positive enable pulse. (author)
Singh, Chandralekha
2009-07-01
One finding of cognitive research is that people do not automatically acquire usable knowledge by spending lots of time on task. Because students' knowledge hierarchy is more fragmented, "knowledge chunks" are smaller than those of experts. The limited capacity of short term memory makes the cognitive load high during problem solving tasks, leaving few cognitive resources available for meta-cognition. The abstract nature of the laws of physics and the chain of reasoning required to draw meaningful inferences makes these issues critical. In order to help students, it is crucial to consider the difficulty of a problem from the perspective of students. We are developing and evaluating interactive problem-solving tutorials to help students in the introductory physics courses learn effective problem-solving strategies while solidifying physics concepts. The self-paced tutorials can provide guidance and support for a variety of problem solving techniques, and opportunity for knowledge and skill acquisition.
Teaching Creative Problem Solving.
Christensen, Kip W.; Martin, Loren
1992-01-01
Interpersonal and cognitive skills, adaptability, and critical thinking can be developed through problem solving and cooperative learning in technology education. These skills have been identified as significant needs of the workplace as well as for functioning in society. (SK)
International Nuclear Information System (INIS)
Vretenar, M
2014-01-01
The main features of radio-frequency linear accelerators are introduced, reviewing the different types of accelerating structures and presenting the main characteristics aspects of linac beam dynamics
Carr, Joseph
1996-01-01
The linear IC market is large and growing, as is the demand for well trained technicians and engineers who understand how these devices work and how to apply them. Linear Integrated Circuits provides in-depth coverage of the devices and their operation, but not at the expense of practical applications in which linear devices figure prominently. This book is written for a wide readership from FE and first degree students, to hobbyists and professionals.Chapter 1 offers a general introduction that will provide students with the foundations of linear IC technology. From chapter 2 onwa
An inherently parallel method for solving discretized diffusion equations
International Nuclear Information System (INIS)
Eccleston, B.R.; Palmer, T.S.
1999-01-01
A Monte Carlo approach to solving linear systems of equations is being investigated in the context of the solution of discretized diffusion equations. While the technique was originally devised decades ago, changes in computer architectures (namely, massively parallel machines) have driven the authors to revisit this technique. There are a number of potential advantages to this approach: (1) Analog Monte Carlo techniques are inherently parallel; this is not necessarily true to today's more advanced linear equation solvers (multigrid, conjugate gradient, etc.); (2) Some forms of this technique are adaptive in that they allow the user to specify locations in the problem where resolution is of particular importance and to concentrate the work at those locations; and (3) These techniques permit the solution of very large systems of equations in that matrix elements need not be stored. The user could trade calculational speed for storage if elements of the matrix are calculated on the fly. The goal of this study is to compare the parallel performance of Monte Carlo linear solvers to that of a more traditional parallelized linear solver. The authors observe the linear speedup that they expect from the Monte Carlo algorithm, given that there is no domain decomposition to cause significant communication overhead. Overall, PETSc outperforms the Monte Carlo solver for the test problem. The PETSc parallel performance improves with larger numbers of unknowns for a given number of processors. Parallel performance of the Monte Carlo technique is independent of the size of the matrix and the number of processes. They are investigating modifications to the scheme to accommodate matrix problems with positive off-diagonal elements. They are also currently coding an on-the-fly version of the algorithm to investigate the solution of very large linear systems
International Nuclear Information System (INIS)
Sumner, H.M.
1969-03-01
The KDF9/EGDON program ZIP MK 2 is the third of a series of programs for off-line digital computer analysis of dynamic systems: it has been designed specifically to cater for the needs of the design or control engineer in having an input scheme which is minimally computer-oriented. It uses numerical algorithms which are as near fool-proof as the author could discover or devise, and has comprehensive diagnostic sections to help the user in the event of faulty data or machine execution. ZIP MK 2 accepts mathematical models comprising first order linear differential and linear algebraic equations, and from these computes and factorises the transfer functions between specified pairs of output and input variables; if desired, the frequency response may be computed from the computed transfer function. The model input scheme is fully compatible with the frequency response programs FRP MK 1 and MK 2, except that, for ZIP MK 2, transport, or time-delays must be converted by the user to Pade or Bode approximations prior to input. ZIP provides the pole-zero plot, (or complex plane analysis), while FRP provides the frequency response and FIFI the time domain analyses. The pole-zero method of analysis has been little used in the past for complex models, especially where transport delays occur, and one of its primary purposes is as a research tool to investigate the usefulness of this method, for process plant, whether nuclear, chemical or other continuous processes. (author)
Linearization Method and Linear Complexity
Tanaka, Hidema
We focus on the relationship between the linearization method and linear complexity and show that the linearization method is another effective technique for calculating linear complexity. We analyze its effectiveness by comparing with the logic circuit method. We compare the relevant conditions and necessary computational cost with those of the Berlekamp-Massey algorithm and the Games-Chan algorithm. The significant property of a linearization method is that it needs no output sequence from a pseudo-random number generator (PRNG) because it calculates linear complexity using the algebraic expression of its algorithm. When a PRNG has n [bit] stages (registers or internal states), the necessary computational cost is smaller than O(2n). On the other hand, the Berlekamp-Massey algorithm needs O(N2) where N(≅2n) denotes period. Since existing methods calculate using the output sequence, an initial value of PRNG influences a resultant value of linear complexity. Therefore, a linear complexity is generally given as an estimate value. On the other hand, a linearization method calculates from an algorithm of PRNG, it can determine the lower bound of linear complexity.
Linear System of Equations, Matrix Inversion, and Linear Programming Using MS Excel
El-Gebeily, M.; Yushau, B.
2008-01-01
In this note, we demonstrate with illustrations two different ways that MS Excel can be used to solve Linear Systems of Equation, Linear Programming Problems, and Matrix Inversion Problems. The advantage of using MS Excel is its availability and transparency (the user is responsible for most of the details of how a problem is solved). Further, we…
Elementary linear programming with applications
Kolman, Bernard
1995-01-01
Linear programming finds the least expensive way to meet given needs with available resources. Its results are used in every area of engineering and commerce: agriculture, oil refining, banking, and air transport. Authors Kolman and Beck present the basic notions of linear programming and illustrate how they are used to solve important common problems. The software on the included disk leads students step-by-step through the calculations. The Second Edition is completely revised and provides additional review material on linear algebra as well as complete coverage of elementary linear program
Said-Houari, Belkacem
2017-01-01
This self-contained, clearly written textbook on linear algebra is easily accessible for students. It begins with the simple linear equation and generalizes several notions from this equation for the system of linear equations and introduces the main ideas using matrices. It then offers a detailed chapter on determinants and introduces the main ideas with detailed proofs. The third chapter introduces the Euclidean spaces using very simple geometric ideas and discusses various major inequalities and identities. These ideas offer a solid basis for understanding general Hilbert spaces in functional analysis. The following two chapters address general vector spaces, including some rigorous proofs to all the main results, and linear transformation: areas that are ignored or are poorly explained in many textbooks. Chapter 6 introduces the idea of matrices using linear transformation, which is easier to understand than the usual theory of matrices approach. The final two chapters are more advanced, introducing t...
Solving the Linear 1D Thermoelasticity Equations with Pure Delay
Directory of Open Access Journals (Sweden)
Denys Ya. Khusainov
2015-01-01
Full Text Available We propose a system of partial differential equations with a single constant delay τ>0 describing the behavior of a one-dimensional thermoelastic solid occupying a bounded interval of R1. For an initial-boundary value problem associated with this system, we prove a well-posedness result in a certain topology under appropriate regularity conditions on the data. Further, we show the solution of our delayed model to converge to the solution of the classical equations of thermoelasticity as τ→0. Finally, we deduce an explicit solution representation for the delay problem.
Students' errors in solving linear equation word problems: Case ...
African Journals Online (AJOL)
kofi.mereku
the modified Newman Error Hierarchical levels (NEAL), which comprise reading, comprehension, transformation, process skills and encoding errors. The results revealed that majority (60%) of the students attempted most of the questions with a few (2%) arriving at the correct answer which implies students have difficulties ...
An Application of Linear Algebra over Lattices
Directory of Open Access Journals (Sweden)
M. Hosseinyazdi
2008-03-01
Full Text Available In this paper, first we consider L n as a semimodule over a complete bounded distributive lattice L. Then we define the basic concepts of module theory for L n. After that, we proved many similar theorems in linear algebra for the space L n. An application of linear algebra over lattices for solving linear systems, was given
An Application of Linear Algebra over Lattices
M. Hosseinyazdi
2008-01-01
In this paper, first we consider L n as a semimodule over a complete bounded distributive lattice L. Then we define the basic concepts of module theory for L n. After that, we proved many similar theorems in linear algebra for the space L n. An application of linear algebra over lattices for solving linear systems, was given
Investigating Integer Restrictions in Linear Programming
Edwards, Thomas G.; Chelst, Kenneth R.; Principato, Angela M.; Wilhelm, Thad L.
2015-01-01
Linear programming (LP) is an application of graphing linear systems that appears in many Algebra 2 textbooks. Although not explicitly mentioned in the Common Core State Standards for Mathematics, linear programming blends seamlessly into modeling with mathematics, the fourth Standard for Mathematical Practice (CCSSI 2010, p. 7). In solving a…
Stoll, R R
1968-01-01
Linear Algebra is intended to be used as a text for a one-semester course in linear algebra at the undergraduate level. The treatment of the subject will be both useful to students of mathematics and those interested primarily in applications of the theory. The major prerequisite for mastering the material is the readiness of the student to reason abstractly. Specifically, this calls for an understanding of the fact that axioms are assumptions and that theorems are logical consequences of one or more axioms. Familiarity with calculus and linear differential equations is required for understand
DEFF Research Database (Denmark)
Chemi, Tatiana
2016-01-01
This chapter aims to deconstruct some persistent myths about creativity: the myth of individualism and of the genius. By looking at literature that approaches creativity as a participatory and distributed phenomenon and by bringing empirical evidence from artists’ studios, the author presents a p......, what can educators at higher education learn from the ways creative groups solve problems? How can artists contribute to inspiring higher education?......This chapter aims to deconstruct some persistent myths about creativity: the myth of individualism and of the genius. By looking at literature that approaches creativity as a participatory and distributed phenomenon and by bringing empirical evidence from artists’ studios, the author presents...... a perspective that is relevant to higher education. The focus here is on how artists solve problems in distributed paths, and on the elements of creative collaboration. Creative problem-solving will be looked at as an ongoing dialogue that artists engage with themselves, with others, with recipients...
Barone, F.; Giordano, G.; Acernese, F.; Romano, R.
2017-10-01
In this paper we present monolithic implementations of tunable mechanical seismometers and accelerometers (horizontal, vertical and angular) based on the UNISA Folded Pendulum configuration, characterized by large measurement band 10-7 ÷ 1 kHz, sensitivity down to ≈ 10-15 m/√ Hz, high directivity > 104 , low weight engineering, seismology, geophysics, civil engineering (buildings, bridges, dams, etc.), space (inertial guide).
Masuda, Y; Misztal, I; Tsuruta, S; Legarra, A; Aguilar, I; Lourenco, D A L; Fragomeni, B O; Lawlor, T J
2016-03-01
The objectives of this study were to develop and evaluate an efficient implementation in the computation of the inverse of genomic relationship matrix with the recursion algorithm, called the algorithm for proven and young (APY), in single-step genomic BLUP. We validated genomic predictions for young bulls with more than 500,000 genotyped animals in final score for US Holsteins. Phenotypic data included 11,626,576 final scores on 7,093,380 US Holstein cows, and genotypes were available for 569,404 animals. Daughter deviations for young bulls with no classified daughters in 2009, but at least 30 classified daughters in 2014 were computed using all the phenotypic data. Genomic predictions for the same bulls were calculated with single-step genomic BLUP using phenotypes up to 2009. We calculated the inverse of the genomic relationship matrix GAPY(-1) based on a direct inversion of genomic relationship matrix on a small subset of genotyped animals (core animals) and extended that information to noncore animals by recursion. We tested several sets of core animals including 9,406 bulls with at least 1 classified daughter, 9,406 bulls and 1,052 classified dams of bulls, 9,406 bulls and 7,422 classified cows, and random samples of 5,000 to 30,000 animals. Validation reliability was assessed by the coefficient of determination from regression of daughter deviation on genomic predictions for the predicted young bulls. The reliabilities were 0.39 with 5,000 randomly chosen core animals, 0.45 with the 9,406 bulls, and 7,422 cows as core animals, and 0.44 with the remaining sets. With phenotypes truncated in 2009 and the preconditioned conjugate gradient to solve mixed model equations, the number of rounds to convergence for core animals defined by bulls was 1,343; defined by bulls and cows, 2,066; and defined by 10,000 random animals, at most 1,629. With complete phenotype data, the number of rounds decreased to 858, 1,299, and at most 1,092, respectively. Setting up GAPY(-1
Solow, Daniel
2014-01-01
This text covers the basic theory and computation for a first course in linear programming, including substantial material on mathematical proof techniques and sophisticated computation methods. Includes Appendix on using Excel. 1984 edition.
Liesen, Jörg
2015-01-01
This self-contained textbook takes a matrix-oriented approach to linear algebra and presents a complete theory, including all details and proofs, culminating in the Jordan canonical form and its proof. Throughout the development, the applicability of the results is highlighted. Additionally, the book presents special topics from applied linear algebra including matrix functions, the singular value decomposition, the Kronecker product and linear matrix equations. The matrix-oriented approach to linear algebra leads to a better intuition and a deeper understanding of the abstract concepts, and therefore simplifies their use in real world applications. Some of these applications are presented in detailed examples. In several ‘MATLAB-Minutes’ students can comprehend the concepts and results using computational experiments. Necessary basics for the use of MATLAB are presented in a short introduction. Students can also actively work with the material and practice their mathematical skills in more than 300 exerc...
Berberian, Sterling K
2014-01-01
Introductory treatment covers basic theory of vector spaces and linear maps - dimension, determinants, eigenvalues, and eigenvectors - plus more advanced topics such as the study of canonical forms for matrices. 1992 edition.
Searle, Shayle R
2012-01-01
This 1971 classic on linear models is once again available--as a Wiley Classics Library Edition. It features material that can be understood by any statistician who understands matrix algebra and basic statistical methods.
Christofilos, N.C.; Polk, I.J.
1959-02-17
Improvements in linear particle accelerators are described. A drift tube system for a linear ion accelerator reduces gap capacity between adjacent drift tube ends. This is accomplished by reducing the ratio of the diameter of the drift tube to the diameter of the resonant cavity. Concentration of magnetic field intensity at the longitudinal midpoint of the external sunface of each drift tube is reduced by increasing the external drift tube diameter at the longitudinal center region.
Systematic Problem Solving in Production: The NAX Approach
DEFF Research Database (Denmark)
Axelsdottir, Aslaug; Nygaard, Martin; Edwards, Kasper
2017-01-01
This paper outlines the NAX problem solving approach developed by a group of problem solving experts at a large Danish Producer of medical equipment. The company, “Medicmeter” is one of Denmark’s leading companies when it comes to lean and it has developed a strong problem solving culture. The ma...
Directory of Open Access Journals (Sweden)
Julie Vercelloni
Full Text Available Recently, attempts to improve decision making in species management have focussed on uncertainties associated with modelling temporal fluctuations in populations. Reducing model uncertainty is challenging; while larger samples improve estimation of species trajectories and reduce statistical errors, they typically amplify variability in observed trajectories. In particular, traditional modelling approaches aimed at estimating population trajectories usually do not account well for nonlinearities and uncertainties associated with multi-scale observations characteristic of large spatio-temporal surveys. We present a Bayesian semi-parametric hierarchical model for simultaneously quantifying uncertainties associated with model structure and parameters, and scale-specific variability over time. We estimate uncertainty across a four-tiered spatial hierarchy of coral cover from the Great Barrier Reef. Coral variability is well described; however, our results show that, in the absence of additional model specifications, conclusions regarding coral trajectories become highly uncertain when considering multiple reefs, suggesting that management should focus more at the scale of individual reefs. The approach presented facilitates the description and estimation of population trajectories and associated uncertainties when variability cannot be attributed to specific causes and origins. We argue that our model can unlock value contained in large-scale datasets, provide guidance for understanding sources of uncertainty, and support better informed decision making.
Introspection in Problem Solving
Jäkel, Frank; Schreiber, Cornell
2013-01-01
Problem solving research has encountered an impasse. Since the seminal work of Newell und Simon (1972) researchers do not seem to have made much theoretical progress (Batchelder and Alexander, 2012; Ohlsson, 2012). In this paper we argue that one factor that is holding back the field is the widespread rejection of introspection among cognitive…
Greene, Kim; Heyck-Williams, Jeff; Timpson Gray, Elicia
2017-01-01
Problem solving spans all grade levels and content areas, as evidenced by this compilation of projects from schools across the United States. In one project, high school girls built a solar-powered tent to serve their city's homeless population. In another project, 4th graders explored historic Jamestown to learn about the voices lost to history.…
Utomo, P.H.; Makarim, R.H.
2017-01-01
A Binary puzzle is a Sudoku-like puzzle with values in each cell taken from the set {0,1} {0,1}. Let n≥4 be an even integer, a solved binary puzzle is an n×n binary array that satisfies the following conditions: (1) no three consecutive ones and no three consecutive zeros in each row and each
Ayrinhac, Simon
2014-01-01
We present in this work a demonstration of the maze-solving problem with electricity. Electric current flowing in a maze as a printed circuit produces Joule heating and the right way is instantaneously revealed with infrared thermal imaging. The basic properties of electric current can be discussed in this context, with this challenging question:…
Toward Solving the Problem of Problem Solving: An Analysis Framework
Roesler, Rebecca A.
2016-01-01
Teaching is replete with problem solving. Problem solving as a skill, however, is seldom addressed directly within music teacher education curricula, and research in music education has not examined problem solving systematically. A framework detailing problem-solving component skills would provide a needed foundation. I observed problem solving…
International Nuclear Information System (INIS)
Alcaraz, J.
2001-01-01
After several years of study e''+ e''- linear colliders in the TeV range have emerged as the major and optimal high-energy physics projects for the post-LHC era. These notes summarize the present status form the main accelerator and detector features to their physics potential. The LHC era. These notes summarize the present status, from the main accelerator and detector features to their physics potential. The LHC is expected to provide first discoveries in the new energy domain, whereas an e''+ e''- linear collider in the 500 GeV-1 TeV will be able to complement it to an unprecedented level of precision in any possible areas: Higgs, signals beyond the SM and electroweak measurements. It is evident that the Linear Collider program will constitute a major step in the understanding of the nature of the new physics beyond the Standard Model. (Author) 22 refs
A Direct Heuristic Algorithm for Linear Programming
Indian Academy of Sciences (India)
Abstract. An (3) mathematically non-iterative heuristic procedure that needs no artificial variable is presented for solving linear programming problems. An optimality test is included. Numerical experiments depict the utility/scope of such a procedure.
Conjugate gradient type methods for linear systems with complex symmetric coefficient matrices
Freund, Roland
1989-01-01
We consider conjugate gradient type methods for the solution of large sparse linear system Ax equals b with complex symmetric coefficient matrices A equals A(T). Such linear systems arise in important applications, such as the numerical solution of the complex Helmholtz equation. Furthermore, most complex non-Hermitian linear systems which occur in practice are actually complex symmetric. We investigate conjugate gradient type iterations which are based on a variant of the nonsymmetric Lanczos algorithm for complex symmetric matrices. We propose a new approach with iterates defined by a quasi-minimal residual property. The resulting algorithm presents several advantages over the standard biconjugate gradient method. We also include some remarks on the obvious approach to general complex linear systems by solving equivalent real linear systems for the real and imaginary parts of x. Finally, numerical experiments for linear systems arising from the complex Helmholtz equation are reported.
Creativity and Problem Solving
DEFF Research Database (Denmark)
Vidal, Rene Victor Valqui
2004-01-01
This paper presents some modern and interdisciplinary concepts about creativity and creative processes of special relevance for Operational Research workers. Central publications in the area Creativity-Operational Research are shortly reviewed. Some creative tools and the Creative Problem Solving...... approach are also discussed. Finally, some applications of these concepts and tools are outlined. Some central references are presented for further study of themes related to creativity or creative tools....
Creativity and problem Solving
Directory of Open Access Journals (Sweden)
René Victor Valqui Vidal
2004-12-01
Full Text Available This paper presents some modern and interdisciplinary concepts about creativity and creative processes of special relevance for Operational Research workers. Central publications in the area Creativity-Operational Research are shortly reviewed. Some creative tools and the Creative Problem Solving approach are also discussed. Finally, some applications of these concepts and tools are outlined. Some central references are presented for further study of themes related to creativity or creative tools.
Generalised Assignment Matrix Methodology in Linear Programming
Jerome, Lawrence
2012-01-01
Discrete Mathematics instructors and students have long been struggling with various labelling and scanning algorithms for solving many important problems. This paper shows how to solve a wide variety of Discrete Mathematics and OR problems using assignment matrices and linear programming, specifically using Excel Solvers although the same…
Linear and non-linear calculations of the hose instability in the ion-focused regime
International Nuclear Information System (INIS)
Buchanan, H.L.
1982-01-01
A simple model is adopted to study the hose instability of an intense relativistic electron beam in a partially neutralized, low density ion channel (ion focused regime). Equations of motion for the beam and the channel are derived and linearized to obtain an approximate dispersion relation. The non-linear equations of motion are then solved numerically and the results compared to linearized data
Generalized Linear Covariance Analysis
Carpenter, James R.; Markley, F. Landis
2014-01-01
This talk presents a comprehensive approach to filter modeling for generalized covariance analysis of both batch least-squares and sequential estimators. We review and extend in two directions the results of prior work that allowed for partitioning of the state space into solve-for'' and consider'' parameters, accounted for differences between the formal values and the true values of the measurement noise, process noise, and textita priori solve-for and consider covariances, and explicitly partitioned the errors into subspaces containing only the influence of the measurement noise, process noise, and solve-for and consider covariances. In this work, we explicitly add sensitivity analysis to this prior work, and relax an implicit assumption that the batch estimator's epoch time occurs prior to the definitive span. We also apply the method to an integrated orbit and attitude problem, in which gyro and accelerometer errors, though not estimated, influence the orbit determination performance. We illustrate our results using two graphical presentations, which we call the variance sandpile'' and the sensitivity mosaic,'' and we compare the linear covariance results to confidence intervals associated with ensemble statistics from a Monte Carlo analysis.
A New Spectral Local Linearization Method for Nonlinear Boundary Layer Flow Problems
Directory of Open Access Journals (Sweden)
S. S. Motsa
2013-01-01
Full Text Available We propose a simple and efficient method for solving highly nonlinear systems of boundary layer flow problems with exponentially decaying profiles. The algorithm of the proposed method is based on an innovative idea of linearizing and decoupling the governing systems of equations and reducing them into a sequence of subsystems of differential equations which are solved using spectral collocation methods. The applicability of the proposed method, hereinafter referred to as the spectral local linearization method (SLLM, is tested on some well-known boundary layer flow equations. The numerical results presented in this investigation indicate that the proposed method, despite being easy to develop and numerically implement, is very robust in that it converges rapidly to yield accurate results and is more efficient in solving very large systems of nonlinear boundary value problems of the similarity variable boundary layer type. The accuracy and numerical stability of the SLLM can further be improved by using successive overrelaxation techniques.
A review on application of neural networks and fuzzy logic to solve hydrothermal scheduling problem
International Nuclear Information System (INIS)
Haroon, S.; Malik, T.N.; Zafar, S.
2014-01-01
Electrical power system is highly complicated having hydro and thermal mix with large number of machines. To reduce power production cost, hydro and thermal resources are mixed. Hydrothermal scheduling is the optimal coordination of hydro and thermal plants to meet the system load demand at minimum possible operational cost while satisfying the system constraints. Hydrothermal scheduling is dynamic, large scale, non-linear and non-convex optimization problem. The classical techniques have failed in solving such problem. Artificial Intelligence Tools based techniques are used now a day to solve this complex optimization problem because of their no requirements on the nature of the problem. The aim of this research paper is to provide a comprehensive survey of literature related to both Artificial Neural Network (ANN) and Fuzzy Logic (FL) as effective optimization algorithms for the hydrothermal scheduling problem. The outcomes along with the merits and demerits of individual techniques are also discussed. (author)
Karloff, Howard
1991-01-01
To this reviewer’s knowledge, this is the first book accessible to the upper division undergraduate or beginning graduate student that surveys linear programming from the Simplex Method…via the Ellipsoid algorithm to Karmarkar’s algorithm. Moreover, its point of view is algorithmic and thus it provides both a history and a case history of work in complexity theory. The presentation is admirable; Karloff's style is informal (even humorous at times) without sacrificing anything necessary for understanding. Diagrams (including horizontal brackets that group terms) aid in providing clarity. The end-of-chapter notes are helpful...Recommended highly for acquisition, since it is not only a textbook, but can also be used for independent reading and study. —Choice Reviews The reader will be well served by reading the monograph from cover to cover. The author succeeds in providing a concise, readable, understandable introduction to modern linear programming. —Mathematics of Computing This is a textbook intend...
Removal of round off errors in the matrix exponential method for solving the heavy nuclide chain
International Nuclear Information System (INIS)
Lee, Hyun Chul; Noh, Jae Man; Joo, Hyung Kook
2005-01-01
Many nodal codes for core simulation adopt the micro-depletion procedure for the depletion analysis. Unlike the macro-depletion procedure, the microdepletion procedure uses micro-cross sections and number densities of important nuclides to generate the macro cross section of a spatial calculational node. Therefore, it needs to solve the chain equations of the nuclides of interest to obtain their number densities. There are several methods such as the matrix exponential method (MEM) and the chain linearization method (CLM) for solving the nuclide chain equations. The former solves chain equations exactly even when the cycles that come from the alpha decay exist in the chain while the latter solves the chain approximately when the cycles exist in the chain. The former has another advantage over the latter. Many nodal codes for depletion analysis, such as MASTER, solve only the hard coded nuclide chains with the CLM. Therefore, if we want to extend the chain by adding some more nuclides to the chain, we have to modify the source code. In contrast, we can extend the chain just by modifying the input in the MEM because it is easy to implement the MEM solver for solving an arbitrary nuclide chain. In spite of these advantages of the MEM, many nodal codes adopt the chain linearization because the former has a large round off error when the flux level is very high or short lived or strong absorber nuclides exist in the chain. In this paper, we propose a new technique to remove the round off errors in the MEM and we compared the performance of the two methods
DEFF Research Database (Denmark)
Hansen, David
2012-01-01
Many industrial production work systems have increased in complexity, and their new business model scompete on innovation, rather than low cost.At a medical device production facility committed to Lean Production, a research project was carried out to use Appreciative Inquiry to better engage...... employee strengths in continuou simprovements of the work system. The research question was: “How can Lean problem solving and Appreciative Inquiry be combined for optimized work system innovation?” The research project was carried out as a co-creation process with close cooperation between researcher...
DEFF Research Database (Denmark)
Foss, Kirsten; Foss, Nicolai Juul
2006-01-01
as a general approach to problem solving. We apply these Simonian ideas to organisational issues, specifically new organisational forms. Specifically, Simonian ideas allow us to develop a morphology of new organisational forms and to point to some design problems that characterise these forms.......Two of Herbert Simon's best-known papers are 'The Architecture of Complexity' and 'The Structure of Ill-Structured Problems.' We discuss the neglected links between these two papers, highlighting the role of decomposition in the context of problems on which constraints have been imposed...
1982-10-01
Artificial Intelig ~ence (Vol. III, edited by Paul R. Cohen and’ Edward A.. Feigenbaum)’, The chapter was written B’ Paul Cohen, with contributions... Artificial Intelligence (Vol. III, edited by Paul R. Cohen and EdWard A. Feigenbaum). The chapter was written by Paul R. Cohen, with contributions by Stephen...Wheevoats"EntermdI’ Planning and Problem ’Solving by Paul R. Cohen Chaptb-rXV-of Volumec III’of the Handbook of Artificial Intelligence edited by Paul R
Solving Differential Equations in R: Package deSolve
In this paper we present the R package deSolve to solve initial value problems (IVP) written as ordinary differential equations (ODE), differential algebraic equations (DAE) of index 0 or 1 and partial differential equations (PDE), the latter solved using the method of lines appr...
Solving Differential Equations in R: Package deSolve
Soetaert, K.E.R.; Petzoldt, T.; Setzer, R.W.
2010-01-01
In this paper we present the R package deSolve to solve initial value problems (IVP) written as ordinary differential equations (ODE), differential algebraic equations (DAE) of index 0 or 1 and partial differential equations (PDE), the latter solved using the method of lines approach. The
On the linear programming bound for linear Lee codes.
Astola, Helena; Tabus, Ioan
2016-01-01
Based on an invariance-type property of the Lee-compositions of a linear Lee code, additional equality constraints can be introduced to the linear programming problem of linear Lee codes. In this paper, we formulate this property in terms of an action of the multiplicative group of the field [Formula: see text] on the set of Lee-compositions. We show some useful properties of certain sums of Lee-numbers, which are the eigenvalues of the Lee association scheme, appearing in the linear programming problem of linear Lee codes. Using the additional equality constraints, we formulate the linear programming problem of linear Lee codes in a very compact form, leading to a fast execution, which allows to efficiently compute the bounds for large parameter values of the linear codes.
Solution of single linear tridiagonal systems and vectorization of the ICCG algorithm on the Cray 1
International Nuclear Information System (INIS)
Kershaw, D.S.
1981-01-01
The numerical algorithms used to solve the physics equation in codes which model laser fusion are examined, it is found that a large number of subroutines require the solution of tridiagonal linear systems of equations. One dimensional radiation transport, thermal and suprathermal electron transport, ion thermal conduction, charged particle and neutron transport, all require the solution of tridiagonal systems of equations. The standard algorithm that has been used in the past on CDC 7600's will not vectorize and so cannot take advantage of the large speed increases possible on the Cray-1 through vectorization. There is however, an alternate algorithm for solving tridiagonal systems, called cyclic reduction, which allows for vectorization, and which is optimal for the Cray-1. Software based on this algorithm is now being used in LASNEX to solve tridiagonal linear systems in the subroutines mentioned above. The new algorithm runs as much as five times faster than the standard algorithm on the Cray-1. The ICCG method is being used to solve the diffusion equation with a nine-point coupling scheme on the CDC 7600. In going from the CDC 7600 to the Cray-1, a large part of the algorithm consists of solving tridiagonal linear systems on each L line of the Lagrangian mesh in a manner which is not vectorizable. An alternate ICCG algorithm for the Cray-1 was developed which utilizes a block form of the cyclic reduction algorithm. This new algorithm allows full vectorization and runs as much as five times faster than the old algorithm on the Cray-1. It is now being used in Cray LASNEX to solve the two-dimensional diffusion equation in all the physics subroutines mentioned above
Linear programming computation
PAN, Ping-Qi
2014-01-01
With emphasis on computation, this book is a real breakthrough in the field of LP. In addition to conventional topics, such as the simplex method, duality, and interior-point methods, all deduced in a fresh and clear manner, it introduces the state of the art by highlighting brand-new and advanced results, including efficient pivot rules, Phase-I approaches, reduced simplex methods, deficient-basis methods, face methods, and pivotal interior-point methods. In particular, it covers the determination of the optimal solution set, feasible-point simplex method, decomposition principle for solving large-scale problems, controlled-branch method based on generalized reduced simplex framework for solving integer LP problems.
Joint shape segmentation with linear programming
Huang, Qixing
2011-01-01
We present an approach to segmenting shapes in a heterogenous shape database. Our approach segments the shapes jointly, utilizing features from multiple shapes to improve the segmentation of each. The approach is entirely unsupervised and is based on an integer quadratic programming formulation of the joint segmentation problem. The program optimizes over possible segmentations of individual shapes as well as over possible correspondences between segments from multiple shapes. The integer quadratic program is solved via a linear programming relaxation, using a block coordinate descent procedure that makes the optimization feasible for large databases. We evaluate the presented approach on the Princeton segmentation benchmark and show that joint shape segmentation significantly outperforms single-shape segmentation techniques. © 2011 ACM.
Reduction of Linear Programming to Linear Approximation
Vaserstein, Leonid N.
2006-01-01
It is well known that every Chebyshev linear approximation problem can be reduced to a linear program. In this paper we show that conversely every linear program can be reduced to a Chebyshev linear approximation problem.
Solved problems in electromagnetics
Salazar Bloise, Félix; Bayón Rojo, Ana; Gascón Latasa, Francisco
2017-01-01
This book presents the fundamental concepts of electromagnetism through problems with a brief theoretical introduction at the beginning of each chapter. The present book has a strong didactic character. It explains all the mathematical steps and the theoretical concepts connected with the development of the problem. It guides the reader to understand the employed procedures to learn to solve the exercises independently. The exercises are structured in a similar way: The chapters begin with easy problems increasing progressively in the level of difficulty. This book is written for students of physics and engineering in the framework of the new European Plans of Study for Bachelor and Master and also for tutors and lecturers. .
Solved problems in electrochemistry
International Nuclear Information System (INIS)
Piron, D.L.
2004-01-01
This book presents calculated solutions to problems in fundamental and applied electrochemistry. It uses industrial data to illustrate scientific concepts and scientific knowledge to solve practical problems. It is subdivided into three parts. The first uses modern basic concepts, the second studies the scientific basis for electrode and electrolyte thermodynamics (including E-pH diagrams and the minimum energy involved in transformations) and the kinetics of rate processes (including the energy lost in heat and in parasite reactions). The third part treats larger problems in electrolysis and power generation, as well as in corrosion and its prevention. Each chapter includes three sections: the presentation of useful principles; some twenty problems with their solutions; and, a set of unsolved problems
Polarized electron sources for linear colliders
International Nuclear Information System (INIS)
Clendenin, J.E.; Ecklund, S.D.; Miller, R.H.; Schultz, D.C.; Sheppard, J.C.
1992-07-01
Linear colliders require high peak current beams with low duty factors. Several methods to produce polarized e - beams for accelerators have been developed. The SLC, the first linear collider, utilizes a photocathode gun with a GaAs cathode. Although photocathode sources are probably the only practical alternative for the next generation of linear colliders, several problems remain to be solved, including high voltage breakdown which poisons the cathode, charge limitations that are associated with the condition of the semiconductor cathode, and a relatively low polarization of ≤5O%. Methods to solve or at least greatly reduce the impact of each of these problems are at hand
Sixth SIAM conference on applied linear algebra: Final program and abstracts. Final technical report
Energy Technology Data Exchange (ETDEWEB)
NONE
1997-12-31
Linear algebra plays a central role in mathematics and applications. The analysis and solution of problems from an amazingly wide variety of disciplines depend on the theory and computational techniques of linear algebra. In turn, the diversity of disciplines depending on linear algebra also serves to focus and shape its development. Some problems have special properties (numerical, structural) that can be exploited. Some are simply so large that conventional approaches are impractical. New computer architectures motivate new algorithms, and fresh ways to look at old ones. The pervasive nature of linear algebra in analyzing and solving problems means that people from a wide spectrum--universities, industrial and government laboratories, financial institutions, and many others--share an interest in current developments in linear algebra. This conference aims to bring them together for their mutual benefit. Abstracts of papers presented are included.
A parallel solver for huge dense linear systems
Badia, J. M.; Movilla, J. L.; Climente, J. I.; Castillo, M.; Marqués, M.; Mayo, R.; Quintana-Ortí, E. S.; Planelles, J.
2011-11-01
HDSS (Huge Dense Linear System Solver) is a Fortran Application Programming Interface (API) to facilitate the parallel solution of very large dense systems to scientists and engineers. The API makes use of parallelism to yield an efficient solution of the systems on a wide range of parallel platforms, from clusters of processors to massively parallel multiprocessors. It exploits out-of-core strategies to leverage the secondary memory in order to solve huge linear systems O(100.000). The API is based on the parallel linear algebra library PLAPACK, and on its Out-Of-Core (OOC) extension POOCLAPACK. Both PLAPACK and POOCLAPACK use the Message Passing Interface (MPI) as the communication layer and BLAS to perform the local matrix operations. The API provides a friendly interface to the users, hiding almost all the technical aspects related to the parallel execution of the code and the use of the secondary memory to solve the systems. In particular, the API can automatically select the best way to store and solve the systems, depending of the dimension of the system, the number of processes and the main memory of the platform. Experimental results on several parallel platforms report high performance, reaching more than 1 TFLOP with 64 cores to solve a system with more than 200 000 equations and more than 10 000 right-hand side vectors. New version program summaryProgram title: Huge Dense System Solver (HDSS) Catalogue identifier: AEHU_v1_1 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEHU_v1_1.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 87 062 No. of bytes in distributed program, including test data, etc.: 1 069 110 Distribution format: tar.gz Programming language: Fortran90, C Computer: Parallel architectures: multiprocessors, computer clusters Operating system
An algorithm for the solution of dynamic linear programs
Psiaki, Mark L.
1989-01-01
The algorithm's objective is to efficiently solve Dynamic Linear Programs (DLP) by taking advantage of their special staircase structure. This algorithm constitutes a stepping stone to an improved algorithm for solving Dynamic Quadratic Programs, which, in turn, would make the nonlinear programming method of Successive Quadratic Programs more practical for solving trajectory optimization problems. The ultimate goal is to being trajectory optimization solution speeds into the realm of real-time control. The algorithm exploits the staircase nature of the large constraint matrix of the equality-constrained DLPs encountered when solving inequality-constrained DLPs by an active set approach. A numerically-stable, staircase QL factorization of the staircase constraint matrix is carried out starting from its last rows and columns. The resulting recursion is like the time-varying Riccati equation from multi-stage LQR theory. The resulting factorization increases the efficiency of all of the typical LP solution operations over that of a dense matrix LP code. At the same time numerical stability is ensured. The algorithm also takes advantage of dynamic programming ideas about the cost-to-go by relaxing active pseudo constraints in a backwards sweeping process. This further decreases the cost per update of the LP rank-1 updating procedure, although it may result in more changes of the active set that if pseudo constraints were relaxed in a non-stagewise fashion. The usual stability of closed-loop Linear/Quadratic optimally-controlled systems, if it carries over to strictly linear cost functions, implies that the saving due to reduced factor update effort may outweigh the cost of an increased number of updates. An aerospace example is presented in which a ground-to-ground rocket's distance is maximized. This example demonstrates the applicability of this class of algorithms to aerospace guidance. It also sheds light on the efficacy of the proposed pseudo constraint relaxation
A Newton method for solving continuous multiple material minimum compliance problems
DEFF Research Database (Denmark)
Stolpe, M; Stegmann, Jan
method, one or two linear saddle point systems are solved. These systems involve the Hessian of the objective function, which is both expensive to compute and completely dense. Therefore, the linear algebra is arranged such that the Hessian is not explicitly formed. The main concern is to solve...
A Newton method for solving continuous multiple material minimum compliance problems
DEFF Research Database (Denmark)
Stolpe, Mathias; Stegmann, Jan
2007-01-01
method, one or two linear saddle point systems are solved. These systems involve the Hessian of the objective function, which is both expensive to compute and completely dense. Therefore, the linear algebra is arranged such that the Hessian is not explicitly formed. The main concern is to solve...
Optimal overlapping of waveform relaxation method for linear differential equations
International Nuclear Information System (INIS)
Yamada, Susumu; Ozawa, Kazufumi
2000-01-01
Waveform relaxation (WR) method is extremely suitable for solving large systems of ordinary differential equations (ODEs) on parallel computers, but the convergence of the method is generally slow. In order to accelerate the convergence, the methods which decouple the system into many subsystems with overlaps some of the components between the adjacent subsystems have been proposed. The methods, in general, converge much faster than the ones without overlapping, but the computational cost per iteration becomes larger due to the increase of the dimension of each subsystem. In this research, the convergence of the WR method for solving constant coefficients linear ODEs is investigated and the strategy to determine the number of overlapped components which minimizes the cost of the parallel computations is proposed. Numerical experiments on an SR2201 parallel computer show that the estimated number of the overlapped components by the proposed strategy is reasonable. (author)
Introduction to computational linear algebra
Nassif, Nabil; Erhel, Jocelyne
2015-01-01
Introduction to Computational Linear Algebra introduces the reader with a background in basic mathematics and computer programming to the fundamentals of dense and sparse matrix computations with illustrating examples. The textbook is a synthesis of conceptual and practical topics in ""Matrix Computations."" The book's learning outcomes are twofold: to understand state-of-the-art computational tools to solve matrix computations problems (BLAS primitives, MATLAB® programming) as well as essential mathematical concepts needed to master the topics of numerical linear algebra. It is suitable for s
Numerical method for solving the three-dimensional time-dependent neutron diffusion equation
International Nuclear Information System (INIS)
Khaled, S.M.; Szatmary, Z.
2005-01-01
A numerical time-implicit method has been developed for solving the coupled three-dimensional time-dependent multi-group neutron diffusion and delayed neutron precursor equations. The numerical stability of the implicit computation scheme and the convergence of the iterative associated processes have been evaluated. The computational scheme requires the solution of large linear systems at each time step. For this purpose, the point over-relaxation Gauss-Seidel method was chosen. A new scheme was introduced instead of the usual source iteration scheme. (author)
AI tools in computer based problem solving
Beane, Arthur J.
1988-01-01
The use of computers to solve value oriented, deterministic, algorithmic problems, has evolved a structured life cycle model of the software process. The symbolic processing techniques used, primarily in research, for solving nondeterministic problems, and those for which an algorithmic solution is unknown, have evolved a different model, much less structured. Traditionally, the two approaches have been used completely independently. With the advent of low cost, high performance 32 bit workstations executing identical software with large minicomputers and mainframes, it became possible to begin to merge both models into a single extended model of computer problem solving. The implementation of such an extended model on a VAX family of micro/mini/mainframe systems is described. Examples in both development and deployment of applications involving a blending of AI and traditional techniques are given.
Scalable Newton-Krylov solver for very large power flow problems
Idema, R.; Lahaye, D.J.P.; Vuik, C.; Van der Sluis, L.
2010-01-01
The power flow problem is generally solved by the Newton-Raphson method with a sparse direct solver for the linear system of equations in each iteration. While this works fine for small power flow problems, we will show that for very large problems the direct solver is very slow and we present
Krylov Subspace Methods for Complex Non-Hermitian Linear Systems. Thesis
Freund, Roland W.
1991-01-01
We consider Krylov subspace methods for the solution of large sparse linear systems Ax = b with complex non-Hermitian coefficient matrices. Such linear systems arise in important applications, such as inverse scattering, numerical solution of time-dependent Schrodinger equations, underwater acoustics, eddy current computations, numerical computations in quantum chromodynamics, and numerical conformal mapping. Typically, the resulting coefficient matrices A exhibit special structures, such as complex symmetry, or they are shifted Hermitian matrices. In this paper, we first describe a Krylov subspace approach with iterates defined by a quasi-minimal residual property, the QMR method, for solving general complex non-Hermitian linear systems. Then, we study special Krylov subspace methods designed for the two families of complex symmetric respectively shifted Hermitian linear systems. We also include some results concerning the obvious approach to general complex linear systems by solving equivalent real linear systems for the real and imaginary parts of x. Finally, numerical experiments for linear systems arising from the complex Helmholtz equation are reported.
Linear programming using Matlab
Ploskas, Nikolaos
2017-01-01
This book offers a theoretical and computational presentation of a variety of linear programming algorithms and methods with an emphasis on the revised simplex method and its components. A theoretical background and mathematical formulation is included for each algorithm as well as comprehensive numerical examples and corresponding MATLAB® code. The MATLAB® implementations presented in this book are sophisticated and allow users to find solutions to large-scale benchmark linear programs. Each algorithm is followed by a computational study on benchmark problems that analyze the computational behavior of the presented algorithms. As a solid companion to existing algorithmic-specific literature, this book will be useful to researchers, scientists, mathematical programmers, and students with a basic knowledge of linear algebra and calculus. The clear presentation enables the reader to understand and utilize all components of simplex-type methods, such as presolve techniques, scaling techniques, pivoting ru...
Solving ptychography with a convex relaxation
Horstmeyer, Roarke; Chen, Richard Y.; Ou, Xiaoze; Ames, Brendan; Tropp, Joel A.; Yang, Changhuei
2015-05-01
Ptychography is a powerful computational imaging technique that transforms a collection of low-resolution images into a high-resolution sample reconstruction. Unfortunately, algorithms that currently solve this reconstruction problem lack stability, robustness, and theoretical guarantees. Recently, convex optimization algorithms have improved the accuracy and reliability of several related reconstruction efforts. This paper proposes a convex formulation of the ptychography problem. This formulation has no local minima, it can be solved using a wide range of algorithms, it can incorporate appropriate noise models, and it can include multiple a priori constraints. The paper considers a specific algorithm, based on low-rank factorization, whose runtime and memory usage are near-linear in the size of the output image. Experiments demonstrate that this approach offers a 25% lower background variance on average than alternating projections, the ptychographic reconstruction algorithm that is currently in widespread use.
Directory of Open Access Journals (Sweden)
Tanwiwat Jaikuna
2017-02-01
Full Text Available Purpose: To develop an in-house software program that is able to calculate and generate the biological dose distribution and biological dose volume histogram by physical dose conversion using the linear-quadratic-linear (LQL model. Material and methods : The Isobio software was developed using MATLAB version 2014b to calculate and generate the biological dose distribution and biological dose volume histograms. The physical dose from each voxel in treatment planning was extracted through Computational Environment for Radiotherapy Research (CERR, and the accuracy was verified by the differentiation between the dose volume histogram from CERR and the treatment planning system. An equivalent dose in 2 Gy fraction (EQD2 was calculated using biological effective dose (BED based on the LQL model. The software calculation and the manual calculation were compared for EQD2 verification with pair t-test statistical analysis using IBM SPSS Statistics version 22 (64-bit. Results: Two and three-dimensional biological dose distribution and biological dose volume histogram were displayed correctly by the Isobio software. Different physical doses were found between CERR and treatment planning system (TPS in Oncentra, with 3.33% in high-risk clinical target volume (HR-CTV determined by D90%, 0.56% in the bladder, 1.74% in the rectum when determined by D2cc, and less than 1% in Pinnacle. The difference in the EQD2 between the software calculation and the manual calculation was not significantly different with 0.00% at p-values 0.820, 0.095, and 0.593 for external beam radiation therapy (EBRT and 0.240, 0.320, and 0.849 for brachytherapy (BT in HR-CTV, bladder, and rectum, respectively. Conclusions : The Isobio software is a feasible tool to generate the biological dose distribution and biological dose volume histogram for treatment plan evaluation in both EBRT and BT.
Energy Technology Data Exchange (ETDEWEB)
Emmerik, Emanuel Leonardus van; Steinberger, Johann Michael; Aredes, Mauricio [Coordenacao dos Programas de Pos-Graduacao de Engenharia (PEE/COPPE/UFRJ), RJ (Brazil). Programa de Engenharia Eletrica
2010-07-01
The optimal deployment of wind generation with the hydro generation is being investigated as a viable option to assist in resolving the constraints coming ahead as a consequence of the tendency of recovery in the Brazilian Amazon basin for expansion of generating facilities. It is in the validity of this research that this work is focused. The value is shown of feasibility studies of using water power generation to offset the variability of wind generation when it is deployed on a large scale. Preliminary results are presented for the variability of wind generation at various cycles, the variability of the availability of hydropower. (author)
Stopping test of iterative methods for solving PDE
International Nuclear Information System (INIS)
Wang Bangrong
1991-01-01
In order to assure the accuracy of the numerical solution of the iterative method for solving PDE (partial differential equation), the stopping test is very important. If the coefficient matrix of the system of linear algebraic equations is strictly diagonal dominant or irreducible weakly diagonal dominant, the stopping test formulas of the iterative method for solving PDE is proposed. Several numerical examples are given to illustrate the applications of the stopping test formulas
Variational iteration method for solving coupled-KdV equations
International Nuclear Information System (INIS)
Assas, Laila M.B.
2008-01-01
In this paper, the He's variational iteration method is applied to solve the non-linear coupled-KdV equations. This method is based on the use of Lagrange multipliers for identification of optimal value of a parameter in a functional. This technique provides a sequence of functions which converge to the exact solution of the coupled-KdV equations. This procedure is a powerful tool for solving coupled-KdV equations
A Photon Free Method to Solve Radiation Transport Equations
International Nuclear Information System (INIS)
Chang, B
2006-01-01
The multi-group discrete-ordinate equations of radiation transfer is solved for the first time by Newton's method. It is a photon free method because the photon variables are eliminated from the radiation equations to yield a N group XN direction smaller but equivalent system of equations. The smaller set of equations can be solved more efficiently than the original set of equations. Newton's method is more stable than the Semi-implicit Linear method currently used by conventional radiation codes
A "feasible direction" search for Lineal Programming problem solving
Directory of Open Access Journals (Sweden)
Jaime U Malpica Angarita
2003-07-01
Full Text Available The study presents an approach to solve linear programming problems with no artificial variables. A primal linear minimization problem is standard form and its associated dual linear maximization problem are used. Initially, the dual (or a partial dual program is solved by a "feasible direction" search, where the Karush-Kuhn-Tucker conditions help to verify its optimality and then its feasibility. The "feasible direction" search exploits the characteristics of the convex polyhedron (or prototype formed by the dual program constraints to find a starting point and then follows line segments, whose directions are found in afine subspaces defined by boundary hyperplanes of polyhedral faces, to find next points up to the (an optimal one. Them, the remaining dual constraints not satisfaced at that optimal dual point, if there are any, are handled as nonbasic variables of the primal program, which is to be solved by such "feasible direction" search.
Innovative problem solving by wild spotted hyenas
Benson-Amram, Sarah; Holekamp, Kay E.
2012-01-01
Innovative animals are those able to solve novel problems or invent novel solutions to existing problems. Despite the important ecological and evolutionary consequences of innovation, we still know very little about the traits that vary among individuals within a species to make them more or less innovative. Here we examine innovative problem solving by spotted hyenas (Crocuta crocuta) in their natural habitat, and demonstrate for the first time in a non-human animal that those individuals exhibiting a greater diversity of initial exploratory behaviours are more successful problem solvers. Additionally, as in earlier work, we found that neophobia was a critical inhibitor of problem-solving success. Interestingly, although juveniles and adults were equally successful in solving the problem, juveniles were significantly more diverse in their initial exploratory behaviours, more persistent and less neophobic than were adults. We found no significant effects of social rank or sex on success, the diversity of initial exploratory behaviours, behavioural persistence or neophobia. Our results suggest that the diversity of initial exploratory behaviours, akin to some measures of human creativity, is an important, but largely overlooked, determinant of problem-solving success in non-human animals. PMID:22874748
Blyth, T S
2002-01-01
Most of the introductory courses on linear algebra develop the basic theory of finite dimensional vector spaces, and in so doing relate the notion of a linear mapping to that of a matrix. Generally speaking, such courses culminate in the diagonalisation of certain matrices and the application of this process to various situations. Such is the case, for example, in our previous SUMS volume Basic Linear Algebra. The present text is a continuation of that volume, and has the objective of introducing the reader to more advanced properties of vector spaces and linear mappings, and consequently of matrices. For readers who are not familiar with the contents of Basic Linear Algebra we provide an introductory chapter that consists of a compact summary of the prerequisites for the present volume. In order to consolidate the student's understanding we have included a large num ber of illustrative and worked examples, as well as many exercises that are strategi cally placed throughout the text. Solutions to the ex...
Design and property analysis of a hybrid linear actuator based on shape memory alloy
International Nuclear Information System (INIS)
Zhang, Xiaoguang; Hu, Jinhong; Mao, Shixin; Dong, Erbao; Yang, Jie
2014-01-01
This paper introduces two methods for solving two bottlelike problems regarding the shape memory alloy (SMA) application as actuators. These methods are ‘rotating output,’ which aims to solve the problem of the low working frequency caused by the demand for cool time, and ‘accumulated shifting,’ which solves the problem of difficult-to-obtain output displacements in a large scale. We also introduce a hybrid linear actuator that applies the two methods and achieves both a strong force and an accurate large output displacement while working at a high frequency based on the SMA wires and DC motors. A prototype of this actuator was fabricated and tested to verify the two methods. This hybrid actuator system dynamic model, which was composed of the constitutive model of the SMA, the electrical and heat transfer behavior of the SMA wires and the dynamics of the linear actuation system, was established and discussed. Our study aims to illuminate the application of an SMA in actuators with the proposed methods with regard to its two main problems. An actuator with a high power-weight ratio and the capability to work at a high frequency, as well as accurate linear step displacements in a large scale, is also presented. (paper)
Solving discretely-constrained MPEC problems with applications in electric power markets
International Nuclear Information System (INIS)
Gabriel, Steven A.; Leuthold, Florian U.
2010-01-01
Many of the European energy markets are characterized by dominant players that own a large share of their respective countries' generation capacities. In addition to that, there is a significant lack of cross-border transmission capacity. Combining both facts justifies the assumption that these dominant players are able to influence the market outcome of an internal European energy market due to strategic behavior. In this paper, we present a mathematical formulation in order to solve a Stackelberg game for a network-constrained energy market using integer programming. The strategic player is the Stackelberg leader and the independent system operator (including the decisions of the competitive fringe firms) acts as follower. We assume that there is one strategic player which results in a mathematical program with equilibrium constraints (MPEC). This MPEC is reformulated as mixed-integer linear program (MILP) by using disjunctive constraints and linearization. The MILP formulation gives the opportunity to solve the problems reliably and paves the way to add discrete constraints to the original MPEC formulation which can be used in order to solve discretely-constrained mathematical programs with equilibrium constraints (DC-MPECs). We report computational results for a small illustrative network as well as a stylized Western European grid with realistic data. (author)
Solving discretely-constrained MPEC problems with applications in electric power markets
Energy Technology Data Exchange (ETDEWEB)
Gabriel, Steven A. [1143 Glenn L. Martin Hall, Department of Civil and Environmental Engineering, University of Maryland, College Park, MD 20742-3021 (United States); Leuthold, Florian U. [Chair of Energy Economics and Public Sector Management, Dresden University of Technology, 01069 Dresden (Germany)
2010-01-15
Many of the European energy markets are characterized by dominant players that own a large share of their respective countries' generation capacities. In addition to that, there is a significant lack of cross-border transmission capacity. Combining both facts justifies the assumption that these dominant players are able to influence the market outcome of an internal European energy market due to strategic behavior. In this paper, we present a mathematical formulation in order to solve a Stackelberg game for a network-constrained energy market using integer programming. The strategic player is the Stackelberg leader and the independent system operator (including the decisions of the competitive fringe firms) acts as follower. We assume that there is one strategic player which results in a mathematical program with equilibrium constraints (MPEC). This MPEC is reformulated as mixed-integer linear program (MILP) by using disjunctive constraints and linearization. The MILP formulation gives the opportunity to solve the problems reliably and paves the way to add discrete constraints to the original MPEC formulation which can be used in order to solve discretely-constrained mathematical programs with equilibrium constraints (DC-MPECs). We report computational results for a small illustrative network as well as a stylized Western European grid with realistic data. (author)
Clock Math — a System for Solving SLEs Exactly
Directory of Open Access Journals (Sweden)
Jakub Hladík
2013-01-01
Full Text Available In this paper, we present a GPU-accelerated hybrid system that solves ill-conditioned systems of linear equations exactly. Exactly means without rounding errors due to using integer arithmetics. First, we scale floating-point numbers up to integers, then we solve dozens of SLEs within different modular arithmetics and then we assemble sub-solutions back using the Chinese remainder theorem. This approach effectively bypasses current CPU floating-point limitations. The system is capable of solving Hilbert’s matrix without losing a single bit of precision, and with a significant speedup compared to existing CPU solvers.
Problem Solving and the Development of Expertise in Management.
Lash, Fredrick B.
This study investigated novice and expert problem solving behavior in management to examine the role of domain specific knowledge on problem solving processes. Forty-one middle level marketing managers in a large petrochemical organization provided think aloud protocols in response to two hypothetical management scenarios. Protocol analysis…
Solving PDEs in Python the FEniCS tutorial I
Langtangen, Hans Petter
2016-01-01
This book offers a concise and gentle introduction to finite element programming in Python based on the popular FEniCS software library. Using a series of examples, including the Poisson equation, the equations of linear elasticity, the incompressible Navier–Stokes equations, and systems of nonlinear advection–diffusion–reaction equations, it guides readers through the essential steps to quickly solving a PDE in FEniCS, such as how to define a finite variational problem, how to set boundary conditions, how to solve linear and nonlinear systems, and how to visualize solutions and structure finite element Python programs. This book is open access under a CC BY license.
The research of radar target tracking observed information linear filter method
Chen, Zheng; Zhao, Xuanzhi; Zhang, Wen
2018-05-01
Aiming at the problems of low precision or even precision divergent is caused by nonlinear observation equation in radar target tracking, a new filtering algorithm is proposed in this paper. In this algorithm, local linearization is carried out on the observed data of the distance and angle respectively. Then the kalman filter is performed on the linearized data. After getting filtered data, a mapping operation will provide the posteriori estimation of target state. A large number of simulation results show that this algorithm can solve above problems effectively, and performance is better than the traditional filtering algorithm for nonlinear dynamic systems.
Solving Kepler's equation using implicit functions
Mortari, Daniele; Elipe, Antonio
2014-01-01
A new approach to solve Kepler's equation based on the use of implicit functions is proposed here. First, new upper and lower bounds are derived for two ranges of mean anomaly. These upper and lower bounds initialize a two-step procedure involving the solution of two implicit functions. These two implicit functions, which are non-rational (polynomial) Bézier functions, can be linear or quadratic, depending on the derivatives of the initial bound values. These are new initial bounds that have been compared and proven more accurate than Serafin's bounds. The procedure reaches machine error accuracy with no more that one quadratic and one linear iterations, experienced in the "tough range", where the eccentricity is close to one and the mean anomaly to zero. The proposed method is particularly suitable for space-based applications with limited computational capability.
Behavioral modeling of the dominant dynamics in input-output transfer of linear(ized) circuits
Beelen, T.G.J.; Maten, ter E.J.W.; Sihaloho, H.J.; Eijndhoven, van S.J.L.
2010-01-01
We present a powerful procedure for determining both the dominant dynamics of the inputoutput transfer and the corresponding most influential circuit parameters of a linear(ized) circuit. The procedure consists of several steps in which a specific (sub)problem is solved and its solution is used in
Normal mode analysis for linear resistive magnetohydrodynamics
International Nuclear Information System (INIS)
Kerner, W.; Lerbinger, K.; Gruber, R.; Tsunematsu, T.
1984-10-01
The compressible, resistive MHD equations are linearized around an equilibrium with cylindrical symmetry and solved numerically as a complex eigenvalue problem. This normal mode code allows to solve for very small resistivity eta proportional 10 -10 . The scaling of growthrates and layer width agrees very well with analytical theory. Especially, both the influence of current and pressure on the instabilities is studied in detail; the effect of resistivity on the ideally unstable internal kink is analyzed. (orig.)
Difficulties in Genetics Problem Solving.
Tolman, Richard R.
1982-01-01
Examined problem-solving strategies of 30 high school students as they solved genetics problems. Proposes a new sequence of teaching genetics based on results: meiosis, sex chromosomes, sex determination, sex-linked traits, monohybrid and dihybrid crosses (humans), codominance (humans), and Mendel's pea experiments. (JN)
Problem Solving, Scaffolding and Learning
Lin, Shih-Yin
2012-01-01
Helping students to construct robust understanding of physics concepts and develop good solving skills is a central goal in many physics classrooms. This thesis examine students' problem solving abilities from different perspectives and explores strategies to scaffold students' learning. In studies involving analogical problem solving…
Problem Solving on a Monorail.
Barrow, Lloyd H.; And Others
1994-01-01
This activity was created to address a lack of problem-solving activities for elementary children. A "monorail" activity from the Evening Science Program for K-3 Students and Parents program is presented to illustrate the problem-solving format. Designed for performance at stations by groups of two students. (LZ)
Solving complex fisheries management problems
DEFF Research Database (Denmark)
Petter Johnsen, Jahn; Eliasen, Søren Qvist
2011-01-01
A crucial issue for the new EU common fisheries policy is how to solve the discard problem. Through a study of the institutional set up and the arrangements for solving the discard problem in Denmark, the Faroe Islands, Iceland and Norway, the article identifies the discard problem as related...
Cooper, Melanie M.; Cox, Charles T., Jr.; Nammouz, Minory; Case, Edward; Stevens, Ronald
2008-01-01
Improving students' problem-solving skills is a major goal for most science educators. While a large body of research on problem solving exists, assessment of meaningful problem solving is very difficult, particularly for courses with large numbers of students in which one-on-one interactions are not feasible. We have used a suite of software…
Physics at the SLC [SLAC Linear Collider
International Nuclear Information System (INIS)
Swartz, M.L.
1990-11-01
The SLAC Linear Collider (SLC) was constructed in the years 1983--1987 for two principal reasons: to develop the accelerator physics and technology that are necessary for the construction of future linear electron-positron colliders; and to produce electron-positron collisions at the Z 0 pole and to study the physics of the weak neutral current. To date, the SLC program has been quite successful at achieving the first goal. The machine has produced and collided high energy electron and positron beams of three-micron transverse size. The problems of operating an open geometry detector in an environment that is more akin to those found in fixed-target experiments than in storage rings have largely been solved. As a physics producing venture, the SLC has been less successful than was originally hoped but more successful than is commonly believed. Some of the results that have been produced by the Mark II experiment with a very modest data sample are competitive with those that have been produced with much larger samples by the four LEP collaborations. At the current, time, SLAC is engaged in an ambitious program to upgrade the SLC luminosity and to exploit one of its unique features, a spin polarized electron beam. These lectures are therefore organized into three sections: a brief description of the SLC; a review of the physics results that have been achieved with the Mark II detector; a description of the SLC's future: the realization and use of a polarized electron beam
Analytical study of dynamic aperture for storage ring by using successive linearization method
International Nuclear Information System (INIS)
Yang Jiancheng; Xia Jiawen; Wu Junxia; Xia Guoxing; Liu Wei; Yin Xuejun
2004-01-01
The determination of dynamic aperture is a critical issue in circular accelerator. In this paper, authors solved the equation of motion including non-linear forces by using successive linearization method and got a criterion for the determining of the dynamic aperture of the machine. Applying this criterion, a storage ring with FODO lattice has been studied. The results are agree well with the tracking results in a large range of linear turn (Q). The purpose is to improve our understanding of the mechanisms driving the particle motion in the presence of non-linear forces and got another mechanism driving instability of particle in storage ring-parametric resonance caused by 'fluctuating transfer matrices' at small amplification
New Method for Solving Inductive Electric Fields in the Ionosphere
Vanhamäki, H.
2005-12-01
We present a new method for calculating inductive electric fields in the ionosphere. It is well established that on large scales the ionospheric electric field is a potential field. This is understandable, since the temporal variations of large scale current systems are generally quite slow, in the timescales of several minutes, so inductive effects should be small. However, studies of Alfven wave reflection have indicated that in some situations inductive phenomena could well play a significant role in the reflection process, and thus modify the nature of ionosphere-magnetosphere coupling. The input to our calculation method are the time series of the potential part of the ionospheric electric field together with the Hall and Pedersen conductances. The output is the time series of the induced rotational part of the ionospheric electric field. The calculation method works in the time-domain and can be used with non-uniform, time-dependent conductances. In addition no particular symmetry requirements are imposed on the input potential electric field. The presented method makes use of special non-local vector basis functions called Cartesian Elementary Current Systems (CECS). This vector basis offers a convenient way of representing curl-free and divergence-free parts of 2-dimensional vector fields and makes it possible to solve the induction problem using simple linear algebra. The new calculation method is validated by comparing it with previously published results for Alfven wave reflection from uniformly conducting ionosphere.
Mukherjee, Kanchan Kumar; Kumar, Narendra; Tripathi, Manjul; Oinam, Arun S; Ahuja, Chirag K; Dhandapani, Sivashanmugam; Kapoor, Rakesh; Ghoshal, Sushmita; Kaur, Rupinder; Bhatt, Sandeep
2017-01-01
To evaluate the feasibility, safety and efficacy of dose fractionated gamma knife radiosurgery (DFGKRS) on a daily schedule beyond the linear quadratic (LQ) model, for large volume arteriovenous malformations (AVMs). Between 2012-16, 14 patients of large AVMs (median volume 26.5 cc) unsuitable for surgery or embolization were treated in 2-3 of DFGKRS sessions. The Leksell G frame was kept in situ during the whole procedure. 86% (n = 12) patients had radiologic evidence of bleed, and 43% (n = 6) had presented with a history of seizures. 57% (n = 8) patients received a daily treatment for 3 days and 43% (n = 6) were on an alternate day (2 fractions) regimen. The marginal dose was split into 2 or 3 fractions of the ideal prescription dose of a single fraction of 23-25 Gy. The median follow up period was 35.6 months (8-57 months). In the three-fraction scheme, the marginal dose ranged from 8.9-11.5 Gy, while in the two-fraction scheme, the marginal dose ranged from 11.3-15 Gy at 50% per fraction. Headache (43%, n = 6) was the most common early postoperative complication, which was controlled with short course steroids. Follow up evaluation of at least three years was achieved in seven patients, who have shown complete nidus obliteration in 43% patients while the obliteration has been in the range of 50-99% in rest of the patients. Overall, there was a 67.8% reduction in the AVM volume at 3 years. Nidus obliteration at 3 years showed a significant rank order correlation with the cumulative prescription dose (p 0.95, P value 0.01), with attainment of near-total (more than 95%) obliteration rates beyond 29 Gy of the cumulative prescription dose. No patient receiving a cumulative prescription dose of less than 31 Gy had any severe adverse reaction. In co-variate adjusted ordinal regression, only the cumulative prescription dose had a significant correlation with common terminology criteria for adverse events (CTCAE) severity (P value 0.04), independent of age, AVM volume
Insightful problem solving in an Asian elephant.
Directory of Open Access Journals (Sweden)
Preston Foerder
Full Text Available The "aha" moment or the sudden arrival of the solution to a problem is a common human experience. Spontaneous problem solving without evident trial and error behavior in humans and other animals has been referred to as insight. Surprisingly, elephants, thought to be highly intelligent, have failed to exhibit insightful problem solving in previous cognitive studies. We tested whether three Asian elephants (Elephas maximus would use sticks or other objects to obtain food items placed out-of-reach and overhead. Without prior trial and error behavior, a 7-year-old male Asian elephant showed spontaneous problem solving by moving a large plastic cube, on which he then stood, to acquire the food. In further testing he showed behavioral flexibility, using this technique to reach other items and retrieving the cube from various locations to use as a tool to acquire food. In the cube's absence, he generalized this tool utilization technique to other objects and, when given smaller objects, stacked them in an attempt to reach the food. The elephant's overall behavior was consistent with the definition of insightful problem solving. Previous failures to demonstrate this ability in elephants may have resulted not from a lack of cognitive ability but from the presentation of tasks requiring trunk-held sticks as potential tools, thereby interfering with the trunk's use as a sensory organ to locate the targeted food.
Insightful problem solving in an Asian elephant.
Foerder, Preston; Galloway, Marie; Barthel, Tony; Moore, Donald E; Reiss, Diana
2011-01-01
The "aha" moment or the sudden arrival of the solution to a problem is a common human experience. Spontaneous problem solving without evident trial and error behavior in humans and other animals has been referred to as insight. Surprisingly, elephants, thought to be highly intelligent, have failed to exhibit insightful problem solving in previous cognitive studies. We tested whether three Asian elephants (Elephas maximus) would use sticks or other objects to obtain food items placed out-of-reach and overhead. Without prior trial and error behavior, a 7-year-old male Asian elephant showed spontaneous problem solving by moving a large plastic cube, on which he then stood, to acquire the food. In further testing he showed behavioral flexibility, using this technique to reach other items and retrieving the cube from various locations to use as a tool to acquire food. In the cube's absence, he generalized this tool utilization technique to other objects and, when given smaller objects, stacked them in an attempt to reach the food. The elephant's overall behavior was consistent with the definition of insightful problem solving. Previous failures to demonstrate this ability in elephants may have resulted not from a lack of cognitive ability but from the presentation of tasks requiring trunk-held sticks as potential tools, thereby interfering with the trunk's use as a sensory organ to locate the targeted food.
Linear Algebra and Smarandache Linear Algebra
Vasantha, Kandasamy
2003-01-01
The present book, on Smarandache linear algebra, not only studies the Smarandache analogues of linear algebra and its applications, it also aims to bridge the need for new research topics pertaining to linear algebra, purely in the algebraic sense. We have introduced Smarandache semilinear algebra, Smarandache bilinear algebra and Smarandache anti-linear algebra and their fuzzy equivalents. Moreover, in this book, we have brought out the study of linear algebra and vector spaces over finite p...
Extended linear chain compounds
Linear chain substances span a large cross section of contemporary chemistry ranging from covalent polymers, to organic charge transfer com plexes to nonstoichiometric transition metal coordination complexes. Their commonality, which coalesced intense interest in the theoretical and exper imental solid state physics/chemistry communities, was based on the obser vation that these inorganic and organic polymeric substrates exhibit striking metal-like electrical and optical properties. Exploitation and extension of these systems has led to the systematic study of both the chemistry and physics of highly and poorly conducting linear chain substances. To gain a salient understanding of these complex materials rich in anomalous aniso tropic electrical, optical, magnetic, and mechanical properties, the conver gence of diverse skills and talents was required. The constructive blending of traditionally segregated disciplines such as synthetic and physical organic, inorganic, and polymer chemistry, crystallog...
Multiscale empirical interpolation for solving nonlinear PDEs
Calo, Victor M.
2014-12-01
In this paper, we propose a multiscale empirical interpolation method for solving nonlinear multiscale partial differential equations. The proposed method combines empirical interpolation techniques and local multiscale methods, such as the Generalized Multiscale Finite Element Method (GMsFEM). To solve nonlinear equations, the GMsFEM is used to represent the solution on a coarse grid with multiscale basis functions computed offline. Computing the GMsFEM solution involves calculating the system residuals and Jacobians on the fine grid. We use empirical interpolation concepts to evaluate these residuals and Jacobians of the multiscale system with a computational cost which is proportional to the size of the coarse-scale problem rather than the fully-resolved fine scale one. The empirical interpolation method uses basis functions which are built by sampling the nonlinear function we want to approximate a limited number of times. The coefficients needed for this approximation are computed in the offline stage by inverting an inexpensive linear system. The proposed multiscale empirical interpolation techniques: (1) divide computing the nonlinear function into coarse regions; (2) evaluate contributions of nonlinear functions in each coarse region taking advantage of a reduced-order representation of the solution; and (3) introduce multiscale proper-orthogonal-decomposition techniques to find appropriate interpolation vectors. We demonstrate the effectiveness of the proposed methods on several nonlinear multiscale PDEs that are solved with Newton\\'s methods and fully-implicit time marching schemes. Our numerical results show that the proposed methods provide a robust framework for solving nonlinear multiscale PDEs on a coarse grid with bounded error and significant computational cost reduction.
Solving or resolving inadequate and noisy tomographic systems
Nolet, G.
1985-01-01
Tomography in seismology often leads to underdetermined and inconsistent systems of linear equations. When solving, care must be taken to keep the propagation of data errors under control. In this paper I test the applicability of three types of damped least-squares algorithms to the kind of
A multiobjective approach for solving cooperative n-person games
Energy Technology Data Exchange (ETDEWEB)
Maali, Yashar [Department of Industrial Engineering, Payam-e-Noor University, Tehran (Iran)
2009-11-15
A linear programming model is introduced to solve cooperative games. The solution is always Pareto optimal. It is based on the idea of the core but instead of requiring rationality for all groups, a multiobjective approach is proposed including the importance weights of the players. A case study illustrates the application of this method. (author)
W-algebra for solving problems with fuzzy parameters
Shevlyakov, A. O.; Matveev, M. G.
2018-03-01
A method of solving the problems with fuzzy parameters by means of a special algebraic structure is proposed. The structure defines its operations through operations on real numbers, which simplifies its use. It avoids deficiencies limiting applicability of the other known structures. Examples for solution of a quadratic equation, a system of linear equations and a network planning problem are given.
A numerical method for solving singular De`s
Energy Technology Data Exchange (ETDEWEB)
Mahaver, W.T.
1996-12-31
A numerical method is developed for solving singular differential equations using steepest descent based on weighted Sobolev gradients. The method is demonstrated on a variety of first and second order problems, including linear constrained, unconstrained, and partially constrained first order problems, a nonlinear first order problem with irregular singularity, and two second order variational problems.
Non-Interior Continuation Method for Solving the Monotone Semidefinite Complementarity Problem
International Nuclear Information System (INIS)
Huang, Z.H.; Han, J.
2003-01-01
Recently, Chen and Tseng extended non-interior continuation smoothing methods for solving linear/ nonlinear complementarity problems to semidefinite complementarity problems (SDCP). In this paper we propose a non-interior continuation method for solving the monotone SDCP based on the smoothed Fischer-Burmeister function, which is shown to be globally linearly and locally quadratically convergent under suitable assumptions. Our algorithm needs at most to solve a linear system of equations at each iteration. In addition, in our analysis on global linear convergence of the algorithm, we need not use the assumption that the Frechet derivative of the function involved in the SDCP is Lipschitz continuous. For non-interior continuation/ smoothing methods for solving the nonlinear complementarity problem, such an assumption has been used widely in the literature in order to achieve global linear convergence results of the algorithms
Problem Solving with General Semantics.
Hewson, David
1996-01-01
Discusses how to use general semantics formulations to improve problem solving at home or at work--methods come from the areas of artificial intelligence/computer science, engineering, operations research, and psychology. (PA)
How to solve mathematical problems
Wickelgren, Wayne A
1995-01-01
Seven problem-solving techniques include inference, classification of action sequences, subgoals, contradiction, working backward, relations between problems, and mathematical representation. Also, problems from mathematics, science, and engineering with complete solutions.
Interactive Problem-Solving Interventions
African Journals Online (AJOL)
Frew Demeke Alemu
concerted efforts of unofficial actors to establish unofficial communication ... Frew Demeke Alemu (LLB, LLM in International Human Rights Law from Lund ..... 24 Tamra Pearson d'Estrée (2009), “Problem-Solving Approaches”, (in The SAGE ...
The analysis and design of linear circuits
Thomas, Roland E; Toussaint, Gregory J
2009-01-01
The Analysis and Design of Linear Circuits, 6e gives the reader the opportunity to not only analyze, but also design and evaluate linear circuits as early as possible. The text's abundance of problems, applications, pedagogical tools, and realistic examples helps engineers develop the skills needed to solve problems, design practical alternatives, and choose the best design from several competing solutions. Engineers searching for an accessible introduction to resistance circuits will benefit from this book that emphasizes the early development of engineering judgment.
Tangram solved? Prefrontal cortex activation analysis during geometric problem solving.
Ayaz, Hasan; Shewokis, Patricia A; Izzetoğlu, Meltem; Çakır, Murat P; Onaral, Banu
2012-01-01
Recent neuroimaging studies have implicated prefrontal and parietal cortices for mathematical problem solving. Mental arithmetic tasks have been used extensively to study neural correlates of mathematical reasoning. In the present study we used geometric problem sets (tangram tasks) that require executive planning and visuospatial reasoning without any linguistic representation interference. We used portable optical brain imaging (functional near infrared spectroscopy--fNIR) to monitor hemodynamic changes within anterior prefrontal cortex during tangram tasks. Twelve healthy subjects were asked to solve a series of computerized tangram puzzles and control tasks that required same geometric shape manipulation without problem solving. Total hemoglobin (HbT) concentration changes indicated a significant increase during tangram problem solving in the right hemisphere. Moreover, HbT changes during failed trials (when no solution found) were significantly higher compared to successful trials. These preliminary results suggest that fNIR can be used to assess cortical activation changes induced by geometric problem solving. Since fNIR is safe, wearable and can be used in ecologically valid environments such as classrooms, this neuroimaging tool may help to improve and optimize learning in educational settings.
Ultrasonic Linear Motor with Two Independent Vibrations
Muneishi, Takeshi; Tomikawa, Yoshiro
2004-09-01
We propose a new structure of an ultrasonic linear motor in order to solve the problems of high-power ultrasonic linear motors that drive the XY-stage for electron beam equipment and to expand the application fields of the motor. We pay special attention to the following three points: (1) the vibration in two directions of the ultrasonic linear motor should not influence mutually each other, (2) the vibration in two directions should be divided into the stage traveling direction and the pressing direction of the ultrasonic linear motor, and (3) the rigidity of the stage traveling direction of the ultrasonic linear motor should be increased. As a result, the supporting method of ultrasonic linear motors is simplified. The efficiency of the motor is improved and temperature rise is reduced. The stage position drift is also improved.
Acoustic emission linear pulse holography
International Nuclear Information System (INIS)
Collins, H.D.; Busse, L.J.; Lemon, D.K.
1983-01-01
This paper describes the emission linear pulse holography which produces a chronological linear holographic image of a flaw by utilizing the acoustic energy emitted during crack growth. A thirty two point sampling array is used to construct phase-only linear holograms of simulated acoustic emission sources on large metal plates. The concept behind the AE linear pulse holography is illustrated, and a block diagram of a data acquisition system to implement the concept is given. Array element spacing, synthetic frequency criteria, and lateral depth resolution are specified. A reference timing transducer positioned between the array and the inspection zone and which inititates the time-of-flight measurements is described. The results graphically illustrate the technique using a one-dimensional FFT computer algorithm (ie. linear backward wave) for an AE image reconstruction
International Nuclear Information System (INIS)
Lee, Jin Pyo; Joo, Han Gyu
2010-01-01
In the thermo-fluid analysis code named CUPID, the linear system of pressure equations must be solved in each iteration step. The time for repeatedly solving the linear system can be quite significant because large sparse matrices of Rank more than 50,000 are involved and the diagonal dominance of the system is hardly hold. Therefore parallelization of the linear system solver is essential to reduce the computing time. Meanwhile, Graphics Processing Units (GPU) have been developed as highly parallel, multi-core processors for the global demand of high quality 3D graphics. If a suitable interface is provided, parallelization using GPU can be available to engineering computing. NVIDIA provides a Software Development Kit(SDK) named CUDA(Compute Unified Device Architecture) to code developers so that they can manage GPUs for parallelization using the C language. In this research, we implement parallel routines for the linear system solver using CUDA, and examine the performance of the parallelization. In the next section, we will describe the method of CUDA parallelization for the CUPID code, and then the performance of the CUDA parallelization will be discussed
Solution of systems of linear algebraic equations by the method of summation of divergent series
International Nuclear Information System (INIS)
Kirichenko, G.A.; Korovin, Ya.S.; Khisamutdinov, M.V.; Shmojlov, V.I.
2015-01-01
A method for solving systems of linear algebraic equations has been proposed on the basis on the summation of the corresponding continued fractions. The proposed algorithm for solving systems of linear algebraic equations is classified as direct algorithms providing an exact solution in a finite number of operations. Examples of solving systems of linear algebraic equations have been presented and the effectiveness of the algorithm has been estimated [ru
Paraxial diffractive elements for space-variant linear transforms
Teiwes, Stephan; Schwarzer, Heiko; Gu, Ben-Yuan
1998-06-01
Optical linear transform architectures bear good potential for future developments of very powerful hybrid vision systems and neural network classifiers. The optical modules of such systems could be used as pre-processors to solve complex linear operations at very high speed in order to simplify an electronic data post-processing. However, the applicability of linear optical architectures is strongly connected with the fundamental question of how to implement a specific linear transform by optical means and physical imitations. The large majority of publications on this topic focusses on the optical implementation of space-invariant transforms by the well-known 4f-setup. Only few papers deal with approaches to implement selected space-variant transforms. In this paper, we propose a simple algebraic method to design diffractive elements for an optical architecture in order to realize arbitrary space-variant transforms. The design procedure is based on a digital model of scalar, paraxial wave theory and leads to optimal element transmission functions within the model. Its computational and physical limitations are discussed in terms of complexity measures. Finally, the design procedure is demonstrated by some examples. Firstly, diffractive elements for the realization of different rotation operations are computed and, secondly, a Hough transform element is presented. The correct optical functions of the elements are proved in computer simulation experiments.
Non-linear buckling of an FGM truncated conical shell surrounded by an elastic medium
International Nuclear Information System (INIS)
Sofiyev, A.H.; Kuruoglu, N.
2013-01-01
In this paper, the non-linear buckling of the truncated conical shell made of functionally graded materials (FGMs) surrounded by an elastic medium has been studied using the large deformation theory with von Karman–Donnell-type of kinematic non-linearity. A two-parameter foundation model (Pasternak-type) is used to describe the shell–foundation interaction. The FGM properties are assumed to vary continuously through the thickness direction. The fundamental relations, the modified Donnell type non-linear stability and compatibility equations of the FGM truncated conical shell resting on the Pasternak-type elastic foundation are derived. By using the Superposition and Galerkin methods, the non-linear stability equations for the FGM truncated conical shell is solved. Finally, influences of variations of Winkler foundation stiffness and shear subgrade modulus of the foundation, compositional profiles and shell characteristics on the dimensionless critical non-linear axial load are investigated. The present results are compared with the available data for a special case. -- Highlights: • Nonlinear buckling of FGM conical shell surrounded by elastic medium is studied. • Pasternak foundation model is used to describe the shell–foundation interaction. • Nonlinear basic equations are derived. • Problem is solved by using Superposition and Galerkin methods. • Influences of various parameters on the nonlinear critical load are investigated
PCX, Interior-Point Linear Programming Solver
International Nuclear Information System (INIS)
Czyzyk, J.
2004-01-01
1 - Description of program or function: PCX solves linear programming problems using the Mehrota predictor-corrector interior-point algorithm. PCX can be called as a subroutine or used in stand-alone mode, with data supplied from an MPS file. The software incorporates modules that can be used separately from the linear programming solver, including a pre-solve routine and data structure definitions. 2 - Methods: The Mehrota predictor-corrector method is a primal-dual interior-point method for linear programming. The starting point is determined from a modified least squares heuristic. Linear systems of equations are solved at each interior-point iteration via a sparse Cholesky algorithm native to the code. A pre-solver is incorporated in the code to eliminate inefficiencies in the user's formulation of the problem. 3 - Restriction on the complexity of the problem: There are no size limitations built into the program. The size of problem solved is limited by RAM and swap space on the user's computer
Uzawa method for fuzzy linear system
Ke Wang
2013-01-01
An Uzawa method is presented for solving fuzzy linear systems whose coefficient matrix is crisp and the right-hand side column is arbitrary fuzzy number vector. The explicit iterative scheme is given. The convergence is analyzed with convergence theorems and the optimal parameter is obtained. Numerical examples are given to illustrate the procedure and show the effectiveness and efficiency of the method.
A LINEAR PROGRAMMING ALGORITHM FOR LEAST-COST SCHEDULING
Directory of Open Access Journals (Sweden)
AYMAN H AL-MOMANI
1999-12-01
Full Text Available In this research, some concepts of linear programming and critical path method are reviewed to describe recent modeling structures that have been of great value in analyzing extended planning horizon project time-cost trade-offs problems. A simplified representation of a small project and a linear programming model is formulated to represent this system. Procedures to solve these various problems formulations were cited and the final solution is obtained using LINDO program. The model developed represents many restrictions and management considerations of the project. It could be used by construction managers in a planning stage to explore numerous possible opportunities to the contractor and predict the effect of a decision on the construction to facilitate a preferred operating policy given different management objectives. An implementation using this method is shown to outperform several other techniques and a large class of test problems. Linear programming show that the algorithm is very promising in practice on a wide variety of time-cost trade-offs problems. This method is simple, applicable to a large network, and generates a shorter computational time at low cost, along with an increase in robustness.
LinvPy : a Python package for linear inverse problems
Beaud, Guillaume François Paul
2016-01-01
The goal of this project is to make a Python package including the tau-estimator algorithm to solve linear inverse problems. The package must be distributed, well documented, easy to use and easy to extend for future developers.
Evaluating forest management policies by parametric linear programing
Daniel I. Navon; Richard J. McConnen
1967-01-01
An analytical and simulation technique, parametric linear programing explores alternative conditions and devises an optimal management plan for each condition. Its application in solving policy-decision problems in the management of forest lands is illustrated in an example.
Near-Regular Structure Discovery Using Linear Programming
Huang, Qixing; Guibas, Leonidas J.; Mitra, Niloy J.
2014-01-01
as an optimization and efficiently solve it using linear programming techniques. Our optimization has a discrete aspect, that is, the connectivity relationships among the elements, as well as a continuous aspect, namely the locations of the elements of interest. Both
Menu-Driven Solver Of Linear-Programming Problems
Viterna, L. A.; Ferencz, D.
1992-01-01
Program assists inexperienced user in formulating linear-programming problems. A Linear Program Solver (ALPS) computer program is full-featured LP analysis program. Solves plain linear-programming problems as well as more-complicated mixed-integer and pure-integer programs. Also contains efficient technique for solution of purely binary linear-programming problems. Written entirely in IBM's APL2/PC software, Version 1.01. Packed program contains licensed material, property of IBM (copyright 1988, all rights reserved).
A Global Optimization Algorithm for Sum of Linear Ratios Problem
Yuelin Gao; Siqiao Jin
2013-01-01
We equivalently transform the sum of linear ratios programming problem into bilinear programming problem, then by using the linear characteristics of convex envelope and concave envelope of double variables product function, linear relaxation programming of the bilinear programming problem is given, which can determine the lower bound of the optimal value of original problem. Therefore, a branch and bound algorithm for solving sum of linear ratios programming problem is put forward, and the c...
International Nuclear Information System (INIS)
Kozarov, A.; Petrov, O.; Antonov, J.; Sotirova, S.; Petrova, B.
2006-01-01
The purpose of the linear wind-power generator described in this article is to decrease the following disadvantages of the common wind-powered turbine: 1) large bending and twisting moments to the blades and the shaft, especially when strong winds and turbulence exist; 2) significant values of the natural oscillation period of the construction result in the possibility of occurrence of destroying resonance oscillations; 3) high velocity of the peripheral parts of the rotor creating a danger for birds; 4) difficulties, connected with the installation and the operation on the mountain ridges and passages where the wind energy potential is the largest. The working surfaces of the generator in questions driven by the wind are not connected with a joint shaft but each moves along a railway track with few oscillations. So the sizes of each component are small and their number can be rather large. The mechanical trajectory is not a circle but a closed outline in a vertical plain, which consists of two rectilinear sectors, one above the other, connected in their ends by semi-circumferences. The mechanical energy of each component turns into electrical on the principle of the linear electrical generator. A regulation is provided when the direction of the wind is perpendicular to the route. A possibility of effectiveness is shown through aiming of additional quantities of air to the movable components by static barriers
Computational linear and commutative algebra
Kreuzer, Martin
2016-01-01
This book combines, in a novel and general way, an extensive development of the theory of families of commuting matrices with applications to zero-dimensional commutative rings, primary decompositions and polynomial system solving. It integrates the Linear Algebra of the Third Millennium, developed exclusively here, with classical algorithmic and algebraic techniques. Even the experienced reader will be pleasantly surprised to discover new and unexpected aspects in a variety of subjects including eigenvalues and eigenspaces of linear maps, joint eigenspaces of commuting families of endomorphisms, multiplication maps of zero-dimensional affine algebras, computation of primary decompositions and maximal ideals, and solution of polynomial systems. This book completes a trilogy initiated by the uncharacteristically witty books Computational Commutative Algebra 1 and 2 by the same authors. The material treated here is not available in book form, and much of it is not available at all. The authors continue to prese...
Numerical linear algebra with applications using Matlab
Ford, William
2014-01-01
Designed for those who want to gain a practical knowledge of modern computational techniques for the numerical solution of linear algebra problems, Numerical Linear Algebra with Applications contains all the material necessary for a first year graduate or advanced undergraduate course on numerical linear algebra with numerous applications to engineering and science. With a unified presentation of computation, basic algorithm analysis, and numerical methods to compute solutions, this book is ideal for solving real-world problems. It provides necessary mathematical background information for
Correlated Levy Noise in Linear Dynamical Systems
International Nuclear Information System (INIS)
Srokowski, T.
2011-01-01
Linear dynamical systems, driven by a non-white noise which has the Levy distribution, are analysed. Noise is modelled by a specific stochastic process which is defined by the Langevin equation with a linear force and the Levy distributed symmetric white noise. Correlation properties of the process are discussed. The Fokker-Planck equation driven by that noise is solved. Distributions have the Levy shape and their width, for a given time, is smaller than for processes in the white noise limit. Applicability of the adiabatic approximation in the case of the linear force is discussed. (author)
Decomposition and (importance) sampling techniques for multi-stage stochastic linear programs
Energy Technology Data Exchange (ETDEWEB)
Infanger, G.
1993-11-01
The difficulty of solving large-scale multi-stage stochastic linear programs arises from the sheer number of scenarios associated with numerous stochastic parameters. The number of scenarios grows exponentially with the number of stages and problems get easily out of hand even for very moderate numbers of stochastic parameters per stage. Our method combines dual (Benders) decomposition with Monte Carlo sampling techniques. We employ importance sampling to efficiently obtain accurate estimates of both expected future costs and gradients and right-hand sides of cuts. The method enables us to solve practical large-scale problems with many stages and numerous stochastic parameters per stage. We discuss the theory of sharing and adjusting cuts between different scenarios in a stage. We derive probabilistic lower and upper bounds, where we use importance path sampling for the upper bound estimation. Initial numerical results turned out to be promising.
Solving a Deconvolution Problem in Photon Spectrometry
Aleksandrov, D; Hille, P T; Polichtchouk, B; Kharlov, Y; Sukhorukov, M; Wang, D; Shabratova, G; Demanov, V; Wang, Y; Tveter, T; Faltys, M; Mao, Y; Larsen, D T; Zaporozhets, S; Sibiryak, I; Lovhoiden, G; Potcheptsov, T; Kucheryaev, Y; Basmanov, V; Mares, J; Yanovsky, V; Qvigstad, H; Zenin, A; Nikolaev, S; Siemiarczuk, T; Yuan, X; Cai, X; Redlich, K; Pavlinov, A; Roehrich, D; Manko, V; Deloff, A; Ma, K; Maruyama, Y; Dobrowolski, T; Shigaki, K; Nikulin, S; Wan, R; Mizoguchi, K; Petrov, V; Mueller, H; Ippolitov, M; Liu, L; Sadovsky, S; Stolpovsky, P; Kurashvili, P; Nomokonov, P; Xu, C; Torii, H; Il'kaev, R; Zhang, X; Peresunko, D; Soloviev, A; Vodopyanov, A; Sugitate, T; Ullaland, K; Huang, M; Zhou, D; Nystrand, J; Punin, V; Yin, Z; Batyunya, B; Karadzhev, K; Nazarov, G; Fil'chagin, S; Nazarenko, S; Buskenes, J I; Horaguchi, T; Djuvsland, O; Chuman, F; Senko, V; Alme, J; Wilk, G; Fehlker, D; Vinogradov, Y; Budilov, V; Iwasaki, T; Ilkiv, I; Budnikov, D; Vinogradov, A; Kazantsev, A; Bogolyubsky, M; Lindal, S; Polak, K; Skaali, B; Mamonov, A; Kuryakin, A; Wikne, J; Skjerdal, K
2010-01-01
We solve numerically a deconvolution problem to extract the undisturbed spectrum from the measured distribution contaminated by the finite resolution of the measuring device. A problem of this kind emerges when one wants to infer the momentum distribution of the neutral pions by detecting the it decay photons using the photon spectrometer of the ALICE LHC experiment at CERN {[}1]. The underlying integral equation connecting the sought for pion spectrum and the measured gamma spectrum has been discretized and subsequently reduced to a system of linear algebraic equations. The latter system, however, is known to be ill-posed and must be regularized to obtain a stable solution. This task has been accomplished here by means of the Tikhonov regularization scheme combined with the L-curve method. The resulting pion spectrum is in an excellent quantitative agreement with the pion spectrum obtained from a Monte Carlo simulation. (C) 2010 Elsevier B.V. All rights reserved.
Modeling and Solving the Train Pathing Problem
Directory of Open Access Journals (Sweden)
Chuen-Yih Chen
2009-04-01
Full Text Available In a railroad system, train pathing is concerned with the assignment of trains to links and tracks, and train timetabling allocates time slots to trains. In this paper, we present an optimization heuristic to solve the train pathing and timetabling problem. This heuristic allows the dwell time of trains in a station or link to be dependent on the assigned tracks. It also allows the minimum clearance time between the trains to depend on their relative status. The heuristic generates a number of alternative paths for each train service in the initialization phase. Then it uses a neighborhood search approach to find good feasible combinations of these paths. A linear program is developed to evaluate the quality of each combination that is encountered. Numerical examples are provided.
Solving modified systems with multiple right-hand sides
Energy Technology Data Exchange (ETDEWEB)
Simoncini, V.; Gallopoulos, E. [Univ. of Patras (Greece)
1996-12-31
In this talk we discuss the iterative solution of large linear systems of the form (A + USV{sup H})X = B, where A is an n x n non-Hermitian matrix, USV{sup H} is a rank-r modification of A and B is of rank s with s, r {much_lt} n. We analyze several approaches that exploit the structure of the coefficient matrix so as to solve the systems more efficiently than if one were to apply a non-hermitian solver to the original systems. In the development of procedures, we take into account the presence of both the low-rank modification and the several right-hand sides. Interesting issues connected to this problem originate from the quest for techniques that accelerate the underlying iterative solvers: preconditioning (e.g. inner-outer iteration strategies), domain decomposition, and continuation methods. Experiments are provided to analyze the behavior of the methods depending on the structure of the rectangular matrices. Preconditioning strategies are explored for an efficient implementation on the transformed systems.
Internet computer coaches for introductory physics problem solving
Xu Ryan, Qing
The ability to solve problems in a variety of contexts is becoming increasingly important in our rapidly changing technological society. Problem-solving is a complex process that is important for everyday life and crucial for learning physics. Although there is a great deal of effort to improve student problem solving skills throughout the educational system, national studies have shown that the majority of students emerge from such courses having made little progress toward developing good problem-solving skills. The Physics Education Research Group at the University of Minnesota has been developing Internet computer coaches to help students become more expert-like problem solvers. During the Fall 2011 and Spring 2013 semesters, the coaches were introduced into large sections (200+ students) of the calculus based introductory mechanics course at the University of Minnesota. This dissertation, will address the research background of the project, including the pedagogical design of the coaches and the assessment of problem solving. The methodological framework of conducting experiments will be explained. The data collected from the large-scale experimental studies will be discussed from the following aspects: the usage and usability of these coaches; the usefulness perceived by students; and the usefulness measured by final exam and problem solving rubric. It will also address the implications drawn from this study, including using this data to direct future coach design and difficulties in conducting authentic assessment of problem-solving.
Customer-centered problem solving.
Samelson, Q B
1999-11-01
If there is no single best way to attract new customers and retain current customers, there is surely an easy way to lose them: fail to solve the problems that arise in nearly every buyer-supplier relationship, or solve them in an unsatisfactory manner. Yet, all too frequently, companies do just that. Either we deny that a problem exists, we exert all our efforts to pin the blame elsewhere, or we "Band-Aid" the problem instead of fixing it, almost guaranteeing that we will face it again and again.
DEFF Research Database (Denmark)
Foss, Kirsten; Foss, Nicolai Juul
as a general approach to problem solving. We apply these Simonian ideas to organizational issues, specifically new organizational forms. Specifically, Simonian ideas allow us to develop a morphology of new organizational forms and to point to some design problems that characterize these forms.Keywords: Herbert...... Simon, problem-solving, new organizational forms. JEL Code: D23, D83......Two of Herbert Simon's best-known papers are "The Architecture of Complexity" and "The Structure of Ill-Structured Problems." We discuss the neglected links between these two papers, highlighting the role of decomposition in the context of problems on which constraints have been imposed...
Interactive problem solving using LOGO
Boecker, Heinz-Dieter; Fischer, Gerhard
2014-01-01
This book is unique in that its stress is not on the mastery of a programming language, but on the importance and value of interactive problem solving. The authors focus on several specific interest worlds: mathematics, computer science, artificial intelligence, linguistics, and games; however, their approach can serve as a model that may be applied easily to other fields as well. Those who are interested in symbolic computing will find that Interactive Problem Solving Using LOGO provides a gentle introduction from which one may move on to other, more advanced computational frameworks or more
Inference rule and problem solving
Energy Technology Data Exchange (ETDEWEB)
Goto, S
1982-04-01
Intelligent information processing signifies an opportunity of having man's intellectual activity executed on the computer, in which inference, in place of ordinary calculation, is used as the basic operational mechanism for such an information processing. Many inference rules are derived from syllogisms in formal logic. The problem of programming this inference function is referred to as a problem solving. Although logically inference and problem-solving are in close relation, the calculation ability of current computers is on a low level for inferring. For clarifying the relation between inference and computers, nonmonotonic logic has been considered. The paper deals with the above topics. 16 references.
Linearly constrained minimax optimization
DEFF Research Database (Denmark)
Madsen, Kaj; Schjær-Jacobsen, Hans
1978-01-01
We present an algorithm for nonlinear minimax optimization subject to linear equality and inequality constraints which requires first order partial derivatives. The algorithm is based on successive linear approximations to the functions defining the problem. The resulting linear subproblems...
Using Coaching to Improve the Teaching of Problem Solving to Year 8 Students in Mathematics
Kargas, Christine Anestis; Stephens, Max
2014-01-01
This study investigated how to improve the teaching of problem solving in a large Melbourne secondary school. Coaching was used to support and equip five teachers, some with limited experiences in teaching problem solving, with knowledge and strategies to build up students' problem solving and reasoning skills. The results showed increased…
Applying Groebner bases to solve reduction problems for Feynman integrals
International Nuclear Information System (INIS)
Smirnov, Alexander V.; Smirnov, Vladimir A.
2006-01-01
We describe how Groebner bases can be used to solve the reduction problem for Feynman integrals, i.e. to construct an algorithm that provides the possibility to express a Feynman integral of a given family as a linear combination of some master integrals. Our approach is based on a generalized Buchberger algorithm for constructing Groebner-type bases associated with polynomials of shift operators. We illustrate it through various examples of reduction problems for families of one- and two-loop Feynman integrals. We also solve the reduction problem for a family of integrals contributing to the three-loop static quark potential
Applying Groebner bases to solve reduction problems for Feynman integrals
Energy Technology Data Exchange (ETDEWEB)
Smirnov, Alexander V. [Mechanical and Mathematical Department and Scientific Research Computer Center of Moscow State University, Moscow 119992 (Russian Federation); Smirnov, Vladimir A. [Nuclear Physics Institute of Moscow State University, Moscow 119992 (Russian Federation)
2006-01-15
We describe how Groebner bases can be used to solve the reduction problem for Feynman integrals, i.e. to construct an algorithm that provides the possibility to express a Feynman integral of a given family as a linear combination of some master integrals. Our approach is based on a generalized Buchberger algorithm for constructing Groebner-type bases associated with polynomials of shift operators. We illustrate it through various examples of reduction problems for families of one- and two-loop Feynman integrals. We also solve the reduction problem for a family of integrals contributing to the three-loop static quark potential.