WorldWideScience

Sample records for residual gmres algorithm

  1. A new adaptive GMRES algorithm for achieving high accuracy

    Energy Technology Data Exchange (ETDEWEB)

    Sosonkina, M.; Watson, L.T.; Kapania, R.K. [Virginia Polytechnic Inst., Blacksburg, VA (United States); Walker, H.F. [Utah State Univ., Logan, UT (United States)

    1996-12-31

    GMRES(k) is widely used for solving nonsymmetric linear systems. However, it is inadequate either when it converges only for k close to the problem size or when numerical error in the modified Gram-Schmidt process used in the GMRES orthogonalization phase dramatically affects the algorithm performance. An adaptive version of GMRES (k) which tunes the restart value k based on criteria estimating the GMRES convergence rate for the given problem is proposed here. The essence of the adaptive GMRES strategy is to adapt the parameter k to the problem, similar in spirit to how a variable order ODE algorithm tunes the order k. With FORTRAN 90, which provides pointers and dynamic memory management, dealing with the variable storage requirements implied by varying k is not too difficult. The parameter k can be both increased and decreased-an increase-only strategy is described next followed by pseudocode.

  2. Some observations on weighted GMRES

    KAUST Repository

    Güttel, Stefan

    2014-01-10

    We investigate the convergence of the weighted GMRES method for solving linear systems. Two different weighting variants are compared with unweighted GMRES for three model problems, giving a phenomenological explanation of cases where weighting improves convergence, and a case where weighting has no effect on the convergence. We also present a new alternative implementation of the weighted Arnoldi algorithm which under known circumstances will be favourable in terms of computational complexity. These implementations of weighted GMRES are compared for a large number of examples. We find that weighted GMRES may outperform unweighted GMRES for some problems, but more often this method is not competitive with other Krylov subspace methods like GMRES with deflated restarting or BICGSTAB, in particular when a preconditioner is used. © 2014 Springer Science+Business Media New York.

  3. Some observations on weighted GMRES

    KAUST Repository

    Gü ttel, Stefan; Pestana, Jennifer

    2014-01-01

    We investigate the convergence of the weighted GMRES method for solving linear systems. Two different weighting variants are compared with unweighted GMRES for three model problems, giving a phenomenological explanation of cases where weighting improves convergence, and a case where weighting has no effect on the convergence. We also present a new alternative implementation of the weighted Arnoldi algorithm which under known circumstances will be favourable in terms of computational complexity. These implementations of weighted GMRES are compared for a large number of examples. We find that weighted GMRES may outperform unweighted GMRES for some problems, but more often this method is not competitive with other Krylov subspace methods like GMRES with deflated restarting or BICGSTAB, in particular when a preconditioner is used. © 2014 Springer Science+Business Media New York.

  4. GMRES and integral operators

    Energy Technology Data Exchange (ETDEWEB)

    Kelley, C.T.; Xue, Z.Q. [North Carolina State Univ., Raleigh, NC (United States)

    1994-12-31

    Many discretizations of integral equations and compact fixed point problems are collectively compact and strongly convergent in spaces of continuous functions. These properties not only lead to stable and convergent approximations but also can be used in the construction of fast multilevel algorithms. Recently the GMRES algorithm has become a standard coarse mesh solver. The purpose of this paper is to show how the special properties of integral operators and their approximations are reflected in the performance of the GMRES iteration and how these properties can be used to strengthen the norm in which convergence takes place. The authors illustrate these ideas with composite Gauss rules for integral equations on the unit interval.

  5. A block variant of the GMRES method on massively parallel processors

    Energy Technology Data Exchange (ETDEWEB)

    Li, Guangye [Cray Research, Inc., Eagan, MN (United States)

    1996-12-31

    This paper presents a block variant of the GMRES method for solving general unsymmetric linear systems. This algorithm generates a transformed Hessenberg matrix by solely using block matrix operations and block data communications. It is shown that this algorithm with block size s, denoted by BVGMRES(s,m), is theoretically equivalent to the GMRES(s*m) method. The numerical results show that this algorithm can be more efficient than the standard GMRES method on a cache based single CPU computer with optimized BLAS kernels. Furthermore, the gain in efficiency is more significant on MPPs due to both efficient block operations and efficient block data communications. Our numerical results also show that in comparison to the standard GMRES method, the more PEs that are used on an MPP, the more efficient the BVGMRES(s,m) algorithm is.

  6. Iterative methods for solving Ax=b, GMRES/FOM versus QMR/BiCG

    Energy Technology Data Exchange (ETDEWEB)

    Cullum, J. [IBM Research Division, Yorktown Heights, NY (United States)

    1996-12-31

    We study the convergence of GMRES/FOM and QMR/BiCG methods for solving nonsymmetric Ax=b. We prove that given the results of a BiCG computation on Ax=b, we can obtain a matrix B with the same eigenvalues as A and a vector c such that the residual norms generated by a FOM computation on Bx=c are identical to those generated by the BiCG computations. Using a unitary equivalence for each of these methods, we obtain test problems where we can easily vary certain spectral properties of the matrices. We use these test problems to study the effects of nonnormality on the convergence of GMRES and QMR, to study the effects of eigenvalue outliers on the convergence of QMR, and to compare the convergence of restarted GMRES, QMR, and BiCGSTAB across a family of normal and nonnormal problems. Our GMRES tests on nonnormal test matrices indicate that nonnormality can have unexpected effects upon the residual norm convergence, giving misleading indications of superior convergence over QMR when the error norms for GMRES are not significantly different from those for QMR. Our QMR tests indicate that the convergence of the QMR residual and error norms is influenced predominantly by small and large eigenvalue outliers and by the character, real, complex, or nearly real, of the outliers and the other eigenvalues. In our comparison tests QMR outperformed GMRES(10) and GMRES(20) on both the normal and nonnormal test matrices.

  7. Application of preconditioned GMRES to the numerical solution of the neutron transport equation

    International Nuclear Information System (INIS)

    Patton, B.W.; Holloway, J.P.

    2002-01-01

    The generalized minimal residual (GMRES) method with right preconditioning is examined as an alternative to both standard and accelerated transport sweeps for the iterative solution of the diamond differenced discrete ordinates neutron transport equation. Incomplete factorization (ILU) type preconditioners are used to determine their effectiveness in accelerating GMRES for this application. ILU(τ), which requires the specification of a dropping criteria τ, proves to be a good choice for the types of problems examined in this paper. The combination of ILU(τ) and GMRES is compared with both DSA and unaccelerated transport sweeps for several model problems. It is found that the computational workload of the ILU(τ)-GMRES combination scales nonlinearly with the number of energy groups and quadrature order, making this technique most effective for problems with a small number of groups and discrete ordinates. However, the cost of preconditioner construction can be amortized over several calculations with different source and/or boundary values. Preconditioners built upon standard transport sweep algorithms are also evaluated as to their effectiveness in accelerating the convergence of GMRES. These preconditioners show better scaling with such problem parameters as the scattering ratio, the number of discrete ordinates, and the number of spatial meshes. These sweeps based preconditioners can also be cast in a matrix free form that greatly reduces storage requirements

  8. Right-Hand Side Dependent Bounds for GMRES Applied to Ill-Posed Problems

    KAUST Repository

    Pestana, Jennifer

    2014-01-01

    © IFIP International Federation for Information Processing 2014. In this paper we apply simple GMRES bounds to the nearly singular systems that arise in ill-posed problems. Our bounds depend on the eigenvalues of the coefficient matrix, the right-hand side vector and the nonnormality of the system. The bounds show that GMRES residuals initially decrease, as residual components associated with large eigenvalues are reduced, after which semi-convergence can be expected because of the effects of small eigenvalues.

  9. A New GMRES(m Method for Markov Chains

    Directory of Open Access Journals (Sweden)

    Bing-Yuan Pu

    2013-01-01

    Full Text Available This paper presents a class of new accelerated restarted GMRES method for calculating the stationary probability vector of an irreducible Markov chain. We focus on the mechanism of this new hybrid method by showing how to periodically combine the GMRES and vector extrapolation method into a much efficient one for improving the convergence rate in Markov chain problems. Numerical experiments are carried out to demonstrate the efficiency of our new algorithm on several typical Markov chain problems.

  10. On Investigating GMRES Convergence using Unitary Matrices

    Czech Academy of Sciences Publication Activity Database

    Duintjer Tebbens, Jurjen; Meurant, G.; Sadok, H.; Strakoš, Z.

    2014-01-01

    Roč. 450, 1 June (2014), s. 83-107 ISSN 0024-3795 Grant - others:GA AV ČR(CZ) M100301201; GA MŠk(CZ) LL1202 Institutional support: RVO:67985807 Keywords : GMRES convergence * unitary matrices * unitary spectra * normal matrices * Krylov residual subspace * Schur parameters Subject RIV: BA - General Mathematics Impact factor: 0.939, year: 2014

  11. Properties of Worst-Case GMRES

    Czech Academy of Sciences Publication Activity Database

    Faber, V.; Liesen, J.; Tichý, Petr

    2013-01-01

    Roč. 34, č. 4 (2013), s. 1500-1519 ISSN 0895-4798 R&D Projects: GA ČR GA13-06684S Grant - others:GA AV ČR(CZ) M10041090 Institutional support: RVO:67985807 Keywords : GMRES method * worst-case convergence * ideal GMRES * matrix approximation problems * minmax Subject RIV: BA - General Mathematics Impact factor: 1.806, year: 2013

  12. Effectiveness of various transport synthetic acceleration methods with and without GMRES

    International Nuclear Information System (INIS)

    Chang, J.H.; Adams, M.L.

    2005-01-01

    We explore the effectiveness of three types of transport synthetic acceleration (TSA) methods as stand-alone methods and as pre-conditioners within the GMRES Krylov solver. The three types are β TSA, 'stretched' TSA, and 'stretched and filtered' (SF) TSA. We analyzed the effectiveness of these algorithms using Fourier mode analysis of model two-dimensional problems with periodic boundary conditions, including problems with alternating layers of different materials. The analyses revealed that both β-TSA and stretched TSA can diverge for fairly weak heterogeneities. Performance of SF TSA, even with the optimum filtering parameter, degrades with heterogeneity. However, with GMRES, all TSA methods are convergent. SF TSA with the optimum filtering parameter was the most effective method. Numerical results support our Fourier mode analysis. (authors)

  13. Minimal residual method stronger than polynomial preconditioning

    Energy Technology Data Exchange (ETDEWEB)

    Faber, V.; Joubert, W.; Knill, E. [Los Alamos National Lab., NM (United States)] [and others

    1994-12-31

    Two popular methods for solving symmetric and nonsymmetric systems of equations are the minimal residual method, implemented by algorithms such as GMRES, and polynomial preconditioning methods. In this study results are given on the convergence rates of these methods for various classes of matrices. It is shown that for some matrices, such as normal matrices, the convergence rates for GMRES and for the optimal polynomial preconditioning are the same, and for other matrices such as the upper triangular Toeplitz matrices, it is at least assured that if one method converges then the other must converge. On the other hand, it is shown that matrices exist for which restarted GMRES always converges but any polynomial preconditioning of corresponding degree makes no progress toward the solution for some initial error. The implications of these results for these and other iterative methods are discussed.

  14. Any Admissible Harmonic Ritz Value Set is Possible for GMRES

    Czech Academy of Sciences Publication Activity Database

    Du, K.; Duintjer Tebbens, Jurjen; Meurant, G.

    2017-01-01

    Roč. 47, September 18 (2017), s. 37-56 ISSN 1068-9613 R&D Projects: GA ČR GA13-06684S Institutional support: RVO:67985807 Keywords : Ritz values * harmonic Ritz values * GMRES convergence * prescribed residual norms * FOM convergence Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 0.925, year: 2016 http://etna.mcs.kent.edu/volumes/2011-2020/vol47/abstract.php?vol=47&pages=37-56

  15. Any Admissible Harmonic Ritz Value Set is Possible for GMRES

    Czech Academy of Sciences Publication Activity Database

    Du, K.; Duintjer Tebbens, Jurjen; Meurant, G.

    2017-01-01

    Roč. 47, September 18 (2017), s. 37-56 ISSN 1068-9613 R&D Projects: GA ČR GA13-06684S Institutional support: RVO:67985807 Keywords : Ritz value s * harmonic Ritz value s * GMRES convergence * prescribed residual norms * FOM convergence Subject RIV: BA - General Mathematics OBOR OECD: Applied mathematics Impact factor: 0.925, year: 2016 http://etna.mcs.kent.edu/volumes/2011-2020/vol47/abstract.php?vol=47&pages=37-56

  16. Application of the DSA preconditioned GMRES formalism to the method of characteristics - First results

    International Nuclear Information System (INIS)

    Le Tellier, R.; Hebert, A.

    2004-01-01

    The method of characteristics is well known for its slow convergence; consequently, as it is often done for SN methods, the Generalized Minimal Residual approach (GMRES) has been investigated for its practical implementation and its high reliability. GMRES is one of the most effective Krylov iterative methods to solve large linear systems. Moreover, the system has been 'left preconditioned' with the Algebraic Collapsing Acceleration (ACA) a variant of the Diffusion Synthetic Acceleration (DSA) based on I. Suslov's former works. This paper presents the first numerical results of these methods in 2D geometries with material discontinuities. Indeed, previous investigations have shown a degraded effectiveness of Diffusion Synthetic Accelerations with this kind of geometries. Results are presented for 9 x 9 Cartesian assemblies in terms of the speed of convergence of the inner iterations (fixed source) of the method of characteristics. It shows a significant improvement on the convergence rate. (authors)

  17. Prescribing the behavior of early terminating GMRES and Arnoldi iterations

    Czech Academy of Sciences Publication Activity Database

    Duintjer Tebbens, Jurjen; Meurant, G.

    2014-01-01

    Roč. 65, č. 1 (2014), s. 69-90 ISSN 1017-1398 R&D Projects: GA AV ČR IAA100300802 Grant - others:GA AV ČR(CZ) M100301201 Institutional research plan: CEZ:AV0Z10300504 Keywords : Arnoldi process * early termination * GMRES method * prescribed GMRES convergence * Arnoldi method * prescribed Ritz values Subject RIV: BA - General Mathematics Impact factor: 1.417, year: 2014

  18. Deflation of Eigenvalues for GMRES in Lattice QCD

    International Nuclear Information System (INIS)

    Morgan, Ronald B.; Wilcox, Walter

    2002-01-01

    Versions of GMRES with deflation of eigenvalues are applied to lattice QCD problems. Approximate eigenvectors corresponding to the smallest eigenvalues are generated at the same time that linear equations are solved. The eigenvectors improve convergence for the linear equations, and they help solve other right-hand sides

  19. Research on an efficient preconditioner using GMRES method for the MOC

    International Nuclear Information System (INIS)

    Takeda, Satoshi; Kitada, Takanori; Smith, Michael A.

    2011-01-01

    The modeling accuracy of reactor analysis techniques has improved considerably with the progressive improvements in computational capabilities. The method of characteristics (MOC) solves the neutron transport equation using tracking lines which simulates the neutron paths. The MOC is an accurate calculation method and is becoming a major solver because of the rapid advancement of the computer. In this methodology, the transport equation is discretized into many spatial meshes and energy wise groups. And the discretization generates a large system which needs a lot of computational costs. To reduce computational costs of MOC calculation, we investigate the Generalized Minimal RESidual (GMRES) method as an accelerator and developed an efficient preconditioner for the MOC calculation. The preconditioner we developed was made by simplifying rigorous preconditioner. And the efficiency was verified by comparing the number of iterations which is calculated by one dimensional MOC code

  20. A look-ahead variant of the Lanczos algorithm and its application to the quasi-minimal residual method for non-Hermitian linear systems. Ph.D. Thesis - Massachusetts Inst. of Technology, Aug. 1991

    Science.gov (United States)

    Nachtigal, Noel M.

    1991-01-01

    The Lanczos algorithm can be used both for eigenvalue problems and to solve linear systems. However, when applied to non-Hermitian matrices, the classical Lanczos algorithm is susceptible to breakdowns and potential instabilities. In addition, the biconjugate gradient (BCG) algorithm, which is the natural generalization of the conjugate gradient algorithm to non-Hermitian linear systems, has a second source of breakdowns, independent of the Lanczos breakdowns. Here, we present two new results. We propose an implementation of a look-ahead variant of the Lanczos algorithm which overcomes the breakdowns by skipping over those steps where a breakdown or a near-breakdown would occur. The new algorithm can handle look-ahead steps of any length and requires the same number of matrix-vector products and inner products per step as the classical Lanczos algorithm without look-ahead. Based on the proposed look-ahead Lanczos algorithm, we then present a novel BCG-like approach, the quasi-minimal residual (QMR) method, which avoids the second source of breakdowns in the BCG algorithm. We present details of the new method and discuss some of its properties. In particular, we discuss the relationship between QMR and BCG, showing how one can recover the BCG iterates, when they exist, from the QMR iterates. We also present convergence results for QMR, showing the connection between QMR and the generalized minimal residual (GMRES) algorithm, the optimal method in this class of methods. Finally, we give some numerical examples, both for eigenvalue computations and for non-Hermitian linear systems.

  1. Convergence of Inner-Iteration GMRES Methods for Rank-Deficient Least Squares Problems

    Czech Academy of Sciences Publication Activity Database

    Morikuni, Keiichi; Hayami, K.

    2015-01-01

    Roč. 36, č. 1 (2015), s. 225-250 ISSN 0895-4798 Institutional support: RVO:67985807 Keywords : least squares problem * iterative methods * preconditioner * inner-outer iteration * GMRES method * stationary iterative method * rank-deficient problem Subject RIV: BA - General Mathematics Impact factor: 1.883, year: 2015

  2. By how much can Residual Minimization Accelerate the Convergence of Orthogonal Residual Methods?

    Czech Academy of Sciences Publication Activity Database

    Gutknecht, M. H.; Rozložník, Miroslav

    2001-01-01

    Roč. 27, - (2001), s. 189-213 ISSN 1017-1398 R&D Projects: GA ČR GA201/98/P108 Institutional research plan: AV0Z1030915 Keywords : system of linear algebraic equations * iterative method * Krylov space method * conjugate gradient method * biconjugate gradient method * CG * CGNE * CGNR * CGS * FOM * GMRes * QMR * TFQMR * residual smoothing * MR smoothing * QMR smoothing Subject RIV: BA - General Mathematics Impact factor: 0.438, year: 2001

  3. Any Ritz Value Behavior Is Possible for Arnoldi and for GMRES

    Czech Academy of Sciences Publication Activity Database

    Duintjer Tebbens, Jurjen; Meurant, G.

    2012-01-01

    Roč. 33, č. 3 (2012), s. 958-978 ISSN 0895-4798 R&D Projects: GA AV ČR IAA100300802 Grant - others:GA AV ČR(CZ) M100300901 Institutional research plan: CEZ:AV0Z10300504 Keywords : Ritz values * Arnoldi process * Arnoldi method * GMRES method * prescribed convergence * interlacing properties Subject RIV: BA - General Mathematics Impact factor: 1.342, year: 2012

  4. Tensor-GMRES method for large sparse systems of nonlinear equations

    Science.gov (United States)

    Feng, Dan; Pulliam, Thomas H.

    1994-01-01

    This paper introduces a tensor-Krylov method, the tensor-GMRES method, for large sparse systems of nonlinear equations. This method is a coupling of tensor model formation and solution techniques for nonlinear equations with Krylov subspace projection techniques for unsymmetric systems of linear equations. Traditional tensor methods for nonlinear equations are based on a quadratic model of the nonlinear function, a standard linear model augmented by a simple second order term. These methods are shown to be significantly more efficient than standard methods both on nonsingular problems and on problems where the Jacobian matrix at the solution is singular. A major disadvantage of the traditional tensor methods is that the solution of the tensor model requires the factorization of the Jacobian matrix, which may not be suitable for problems where the Jacobian matrix is large and has a 'bad' sparsity structure for an efficient factorization. We overcome this difficulty by forming and solving the tensor model using an extension of a Newton-GMRES scheme. Like traditional tensor methods, we show that the new tensor method has significant computational advantages over the analogous Newton counterpart. Consistent with Krylov subspace based methods, the new tensor method does not depend on the factorization of the Jacobian matrix. As a matter of fact, the Jacobian matrix is never needed explicitly.

  5. Super-resolution reconstruction of MR image with a novel residual learning network algorithm

    Science.gov (United States)

    Shi, Jun; Liu, Qingping; Wang, Chaofeng; Zhang, Qi; Ying, Shihui; Xu, Haoyu

    2018-04-01

    Spatial resolution is one of the key parameters of magnetic resonance imaging (MRI). The image super-resolution (SR) technique offers an alternative approach to improve the spatial resolution of MRI due to its simplicity. Convolutional neural networks (CNN)-based SR algorithms have achieved state-of-the-art performance, in which the global residual learning (GRL) strategy is now commonly used due to its effectiveness for learning image details for SR. However, the partial loss of image details usually happens in a very deep network due to the degradation problem. In this work, we propose a novel residual learning-based SR algorithm for MRI, which combines both multi-scale GRL and shallow network block-based local residual learning (LRL). The proposed LRL module works effectively in capturing high-frequency details by learning local residuals. One simulated MRI dataset and two real MRI datasets have been used to evaluate our algorithm. The experimental results show that the proposed SR algorithm achieves superior performance to all of the other compared CNN-based SR algorithms in this work.

  6. A Nonlinear GMRES Optimization Algorithm for Canonical Tensor Decomposition

    OpenAIRE

    De Sterck, Hans

    2011-01-01

    A new algorithm is presented for computing a canonical rank-R tensor approximation that has minimal distance to a given tensor in the Frobenius norm, where the canonical rank-R tensor consists of the sum of R rank-one components. Each iteration of the method consists of three steps. In the first step, a tentative new iterate is generated by a stand-alone one-step process, for which we use alternating least squares (ALS). In the second step, an accelerated iterate is generated by a nonlinear g...

  7. Iterative Regularization with Minimum-Residual Methods

    DEFF Research Database (Denmark)

    Jensen, Toke Koldborg; Hansen, Per Christian

    2007-01-01

    subspaces. We provide a combination of theory and numerical examples, and our analysis confirms the experience that MINRES and MR-II can work as general regularization methods. We also demonstrate theoretically and experimentally that the same is not true, in general, for GMRES and RRGMRES their success......We study the regularization properties of iterative minimum-residual methods applied to discrete ill-posed problems. In these methods, the projection onto the underlying Krylov subspace acts as a regularizer, and the emphasis of this work is on the role played by the basis vectors of these Krylov...... as regularization methods is highly problem dependent....

  8. Iterative regularization with minimum-residual methods

    DEFF Research Database (Denmark)

    Jensen, Toke Koldborg; Hansen, Per Christian

    2006-01-01

    subspaces. We provide a combination of theory and numerical examples, and our analysis confirms the experience that MINRES and MR-II can work as general regularization methods. We also demonstrate theoretically and experimentally that the same is not true, in general, for GMRES and RRGMRES - their success......We study the regularization properties of iterative minimum-residual methods applied to discrete ill-posed problems. In these methods, the projection onto the underlying Krylov subspace acts as a regularizer, and the emphasis of this work is on the role played by the basis vectors of these Krylov...... as regularization methods is highly problem dependent....

  9. The performances of R GPU implementations of the GMRES method

    Directory of Open Access Journals (Sweden)

    Bogdan Oancea

    2018-03-01

    Full Text Available Although the performance of commodity computers has improved drastically with the introduction of multicore processors and GPU computing, the standard R distribution is still based on single-threaded model of computation, using only a small fraction of the computational power available now for most desktops and laptops. Modern statistical software packages rely on high performance implementations of the linear algebra routines there are at the core of several important leading edge statistical methods. In this paper we present a GPU implementation of the GMRES iterative method for solving linear systems. We compare the performance of this implementation with a pure single threaded version of the CPU. We also investigate the performance of our implementation using different GPU packages available now for R such as gmatrix, gputools or gpuR which are based on CUDA or OpenCL frameworks.

  10. Deflation of eigenvalues for iterative methods in lattice QCD

    International Nuclear Information System (INIS)

    Darnell, Dean; Morgan, Ronald B.; Wilcox, Walter

    2004-01-01

    Work on generalizing the deflated, restarted GMRES algorithm, useful in lattice studies using stochastic noise methods, is reported. We first show how the multi-mass extension of deflated GMRES can be implemented. We then give a deflated GMRES method that can be used on multiple right-hand sides of Aχ = b in an efficient manner. We also discuss and give numerical results on the possibilty of combining deflated GMRES for the first right hand side with a deflated BiCGStab algorithm for the subsequent right hand sides

  11. A General Algorithm for Reusing Krylov Subspace Information. I. Unsteady Navier-Stokes

    Science.gov (United States)

    Carpenter, Mark H.; Vuik, C.; Lucas, Peter; vanGijzen, Martin; Bijl, Hester

    2010-01-01

    A general algorithm is developed that reuses available information to accelerate the iterative convergence of linear systems with multiple right-hand sides A x = b (sup i), which are commonly encountered in steady or unsteady simulations of nonlinear equations. The algorithm is based on the classical GMRES algorithm with eigenvector enrichment but also includes a Galerkin projection preprocessing step and several novel Krylov subspace reuse strategies. The new approach is applied to a set of test problems, including an unsteady turbulent airfoil, and is shown in some cases to provide significant improvement in computational efficiency relative to baseline approaches.

  12. Investigation of Diesel’s Residual Noise on Predictive Vehicles Noise Cancelling using LMS Adaptive Algorithm

    Science.gov (United States)

    Arttini Dwi Prasetyowati, Sri; Susanto, Adhi; Widihastuti, Ida

    2017-04-01

    Every noise problems require different solution. In this research, the noise that must be cancelled comes from roadway. Least Mean Square (LMS) adaptive is one of the algorithm that can be used to cancel that noise. Residual noise always appears and could not be erased completely. This research aims to know the characteristic of residual noise from vehicle’s noise and analysis so that it is no longer appearing as a problem. LMS algorithm was used to predict the vehicle’s noise and minimize the error. The distribution of the residual noise could be observed to determine the specificity of the residual noise. The statistic of the residual noise close to normal distribution with = 0,0435, = 1,13 and the autocorrelation of the residual noise forming impulse. As a conclusion the residual noise is insignificant.

  13. An efficient preconditioning technique using Krylov subspace methods for 3D characteristics solvers

    International Nuclear Information System (INIS)

    Dahmani, M.; Le Tellier, R.; Roy, R.; Hebert, A.

    2005-01-01

    The Generalized Minimal RESidual (GMRES) method, using a Krylov subspace projection, is adapted and implemented to accelerate a 3D iterative transport solver based on the characteristics method. Another acceleration technique called the self-collision rebalancing technique (SCR) can also be used to accelerate the solution or as a left preconditioner for GMRES. The GMRES method is usually used to solve a linear algebraic system (Ax=b). It uses K(r (o) ,A) as projection subspace and AK(r (o) ,A) for the orthogonalization of the residual. This paper compares the performance of these two combined methods on various problems. To implement the GMRES iterative method, the characteristics equations are derived in linear algebra formalism by using the equivalence between the method of characteristics and the method of collision probability to end up with a linear algebraic system involving fluxes and currents. Numerical results show good performance of the GMRES technique especially for the cases presenting large material heterogeneity with a scattering ratio close to 1. Similarly, the SCR preconditioning slightly increases the GMRES efficiency

  14. SNBRFinder: A Sequence-Based Hybrid Algorithm for Enhanced Prediction of Nucleic Acid-Binding Residues.

    Directory of Open Access Journals (Sweden)

    Xiaoxia Yang

    Full Text Available Protein-nucleic acid interactions are central to various fundamental biological processes. Automated methods capable of reliably identifying DNA- and RNA-binding residues in protein sequence are assuming ever-increasing importance. The majority of current algorithms rely on feature-based prediction, but their accuracy remains to be further improved. Here we propose a sequence-based hybrid algorithm SNBRFinder (Sequence-based Nucleic acid-Binding Residue Finder by merging a feature predictor SNBRFinderF and a template predictor SNBRFinderT. SNBRFinderF was established using the support vector machine whose inputs include sequence profile and other complementary sequence descriptors, while SNBRFinderT was implemented with the sequence alignment algorithm based on profile hidden Markov models to capture the weakly homologous template of query sequence. Experimental results show that SNBRFinderF was clearly superior to the commonly used sequence profile-based predictor and SNBRFinderT can achieve comparable performance to the structure-based template methods. Leveraging the complementary relationship between these two predictors, SNBRFinder reasonably improved the performance of both DNA- and RNA-binding residue predictions. More importantly, the sequence-based hybrid prediction reached competitive performance relative to our previous structure-based counterpart. Our extensive and stringent comparisons show that SNBRFinder has obvious advantages over the existing sequence-based prediction algorithms. The value of our algorithm is highlighted by establishing an easy-to-use web server that is freely accessible at http://ibi.hzau.edu.cn/SNBRFinder.

  15. SNBRFinder: A Sequence-Based Hybrid Algorithm for Enhanced Prediction of Nucleic Acid-Binding Residues.

    Science.gov (United States)

    Yang, Xiaoxia; Wang, Jia; Sun, Jun; Liu, Rong

    2015-01-01

    Protein-nucleic acid interactions are central to various fundamental biological processes. Automated methods capable of reliably identifying DNA- and RNA-binding residues in protein sequence are assuming ever-increasing importance. The majority of current algorithms rely on feature-based prediction, but their accuracy remains to be further improved. Here we propose a sequence-based hybrid algorithm SNBRFinder (Sequence-based Nucleic acid-Binding Residue Finder) by merging a feature predictor SNBRFinderF and a template predictor SNBRFinderT. SNBRFinderF was established using the support vector machine whose inputs include sequence profile and other complementary sequence descriptors, while SNBRFinderT was implemented with the sequence alignment algorithm based on profile hidden Markov models to capture the weakly homologous template of query sequence. Experimental results show that SNBRFinderF was clearly superior to the commonly used sequence profile-based predictor and SNBRFinderT can achieve comparable performance to the structure-based template methods. Leveraging the complementary relationship between these two predictors, SNBRFinder reasonably improved the performance of both DNA- and RNA-binding residue predictions. More importantly, the sequence-based hybrid prediction reached competitive performance relative to our previous structure-based counterpart. Our extensive and stringent comparisons show that SNBRFinder has obvious advantages over the existing sequence-based prediction algorithms. The value of our algorithm is highlighted by establishing an easy-to-use web server that is freely accessible at http://ibi.hzau.edu.cn/SNBRFinder.

  16. Optimization of Residual Stress of High Temperature Treatment Using Genetic Algorithm and Neural Network

    Directory of Open Access Journals (Sweden)

    M. Susmikanti

    2015-12-01

    Full Text Available In a nuclear industry area, high temperature treatment of materials is a factor which requires special attention. Assessment needs to be conducted on the properties of the materials used, including the strength of the materials. The measurement of material properties under thermal processes may reflect residual stresses. The use of Genetic Algorithm (GA to determine the optimal residual stress is one way to determine the strength of a material. In residual stress modeling with several parameters, it is sometimes difficult to solve for the optimal value through analytical or numerical calculations. Here, GA is an efficient algorithm which can generate the optimal values, both minima and maxima. The purposes of this research are to obtain the optimization of variable in residual stress models using GA and to predict the center of residual stress distribution, using fuzzy neural network (FNN while the artificial neural network (ANN used for modeling. In this work a single-material 316/316L stainless steel bar is modeled. The minimal residual stresses of the material at high temperatures were obtained with GA and analytical calculations. At a temperature of 6500C, the GA optimal residual stress estimation converged at –711.3689 MPa at adistance of 0.002934 mm from center point, whereas the analytical calculation result at that temperature and position is -975.556 MPa . At a temperature of 8500C, the GA result was -969.868 MPa at 0.002757 mm from the center point, while with analytical result was -1061.13 MPa. The difference in residual stress between GA and analytical results at a temperatureof6500C is about 27 %, while at 8500C it is 8.67 %. The distribution of residual stress showed a grouping concentrated around a coordinate of (-76; 76 MPa. The residuals stress model is a degree-two polynomial with coefficients of 50.33, -76.54, and -55.2, respectively, with a standard deviation of 7.874.

  17. Efficient stabilization and acceleration of numerical simulation of fluid flows by residual recombination

    Science.gov (United States)

    Citro, V.; Luchini, P.; Giannetti, F.; Auteri, F.

    2017-09-01

    The study of the stability of a dynamical system described by a set of partial differential equations (PDEs) requires the computation of unstable states as the control parameter exceeds its critical threshold. Unfortunately, the discretization of the governing equations, especially for fluid dynamic applications, often leads to very large discrete systems. As a consequence, matrix based methods, like for example the Newton-Raphson algorithm coupled with a direct inversion of the Jacobian matrix, lead to computational costs too large in terms of both memory and execution time. We present a novel iterative algorithm, inspired by Krylov-subspace methods, which is able to compute unstable steady states and/or accelerate the convergence to stable configurations. Our new algorithm is based on the minimization of the residual norm at each iteration step with a projection basis updated at each iteration rather than at periodic restarts like in the classical GMRES method. The algorithm is able to stabilize any dynamical system without increasing the computational time of the original numerical procedure used to solve the governing equations. Moreover, it can be easily inserted into a pre-existing relaxation (integration) procedure with a call to a single black-box subroutine. The procedure is discussed for problems of different sizes, ranging from a small two-dimensional system to a large three-dimensional problem involving the Navier-Stokes equations. We show that the proposed algorithm is able to improve the convergence of existing iterative schemes. In particular, the procedure is applied to the subcritical flow inside a lid-driven cavity. We also discuss the application of Boostconv to compute the unstable steady flow past a fixed circular cylinder (2D) and boundary-layer flow over a hemispherical roughness element (3D) for supercritical values of the Reynolds number. We show that Boostconv can be used effectively with any spatial discretization, be it a finite

  18. Multiple solutions to dense systems in radar scattering using a preconditioned block GMRES solver

    Energy Technology Data Exchange (ETDEWEB)

    Boyse, W.E. [Advanced Software Resources, Inc., Santa Clara, CA (United States)

    1996-12-31

    Multiple right-hand sides occur in radar scattering calculations in the computation of the simulated radar return from a body at a large number of angles. Each desired angle requires a right-hand side vector to be computed and the solution generated. These right-hand sides are naturally smooth functions of the angle parameters and this property is utilized in a novel way to compute solutions an order of magnitude faster than LINPACK The modeling technique addressed is the Method of Moments (MOM), i.e. a boundary element method for time harmonic Maxwell`s equations. Discretization by this method produces general complex dense systems of rank 100`s to 100,000`s. The usual way to produce the required multiple solutions is via LU factorization and solution routines such as found in LINPACK. Our method uses the block GMRES iterative method to directly iterate a subset of the desired solutions to convergence.

  19. Frequency-domain optical tomographic image reconstruction algorithm with the simplified spherical harmonics (SP3) light propagation model.

    Science.gov (United States)

    Kim, Hyun Keol; Montejo, Ludguier D; Jia, Jingfei; Hielscher, Andreas H

    2017-06-01

    We introduce here the finite volume formulation of the frequency-domain simplified spherical harmonics model with n -th order absorption coefficients (FD-SP N ) that approximates the frequency-domain equation of radiative transfer (FD-ERT). We then present the FD-SP N based reconstruction algorithm that recovers absorption and scattering coefficients in biological tissue. The FD-SP N model with 3 rd order absorption coefficient (i.e., FD-SP 3 ) is used as a forward model to solve the inverse problem. The FD-SP 3 is discretized with a node-centered finite volume scheme and solved with a restarted generalized minimum residual (GMRES) algorithm. The absorption and scattering coefficients are retrieved using a limited-memory Broyden-Fletcher-Goldfarb-Shanno (L-BFGS) algorithm. Finally, the forward and inverse algorithms are evaluated using numerical phantoms with optical properties and size that mimic small-volume tissue such as finger joints and small animals. The forward results show that the FD-SP 3 model approximates the FD-ERT (S 12 ) solution within relatively high accuracy; the average error in the phase (<3.7%) and the amplitude (<7.1%) of the partial current at the boundary are reported. From the inverse results we find that the absorption and scattering coefficient maps are more accurately reconstructed with the SP 3 model than those with the SP 1 model. Therefore, this work shows that the FD-SP 3 is an efficient model for optical tomographic imaging of small-volume media with non-diffuse properties both in terms of computational time and accuracy as it requires significantly lower CPU time than the FD-ERT (S 12 ) and also it is more accurate than the FD-SP 1 .

  20. Residual and Backward Error Bounds in Minimum Residual Krylov Subspace Methods

    Czech Academy of Sciences Publication Activity Database

    Paige, C. C.; Strakoš, Zdeněk

    2002-01-01

    Roč. 23, č. 6 (2002), s. 1899-1924 ISSN 1064-8275 R&D Projects: GA AV ČR IAA1030103 Institutional research plan: AV0Z1030915 Keywords : linear equations * eigenproblem * large sparse matrices * iterative solutions * Krylov subspace methods * Arnoldi method * GMRES * modified Gram-Schmidt * least squares * total least squares * singular values Subject RIV: BA - General Mathematics Impact factor: 1.291, year: 2002

  1. Parallel Jacobian-free Newton Krylov solution of the discrete ordinates method with flux limiters for 3D radiative transfer

    International Nuclear Information System (INIS)

    Godoy, William F.; Liu Xu

    2012-01-01

    The present study introduces a parallel Jacobian-free Newton Krylov (JFNK) general minimal residual (GMRES) solution for the discretized radiative transfer equation (RTE) in 3D, absorbing, emitting and scattering media. For the angular and spatial discretization of the RTE, the discrete ordinates method (DOM) and the finite volume method (FVM) including flux limiters are employed, respectively. Instead of forming and storing a large Jacobian matrix, JFNK methods allow for large memory savings as the required Jacobian-vector products are rather approximated by semiexact and numerical formulations, for which convergence and computational times are presented. Parallelization of the GMRES solution is introduced in a combined memory-shared/memory-distributed formulation that takes advantage of the fact that only large vector arrays remain in the JFNK process. Results are presented for 3D test cases including a simple homogeneous, isotropic medium and a more complex non-homogeneous, non-isothermal, absorbing–emitting and anisotropic scattering medium with collimated intensities. Additionally, convergence and stability of Gram–Schmidt and Householder orthogonalizations for the Arnoldi process in the parallel GMRES algorithms are discussed and analyzed. Overall, the introduction of JFNK methods results in a parallel, yet scalable to the tested 2048 processors, and memory affordable solution to 3D radiative transfer problems without compromising the accuracy and convergence of a Newton-like solution.

  2. R3GMRES: including prior information in GMRES-type methods for discrete inverse problems

    DEFF Research Database (Denmark)

    Dong, Yiqiu; Garde, Henrik; Hansen, Per Christian

    2014-01-01

    Lothar Reichel and his collaborators proposed several iterative algorithms that augment the underlying Krylov subspace with an additional low-dimensional subspace in order to produce improved regularized solutions. We take a closer look at this approach and investigate a particular Regularized Ra...

  3. Preconditioned Krylov and Gauss-Seidel solutions of response matrix equations

    International Nuclear Information System (INIS)

    Lewis, E.E.; Smith, M.A.; Yang, W.S.; Wollaber, A.

    2011-01-01

    The use of preconditioned Krylov methods is examined as an alternative to the partitioned matrix acceleration applied to red-black Gauss Seidel (RBGS) iteration that is presently used in the variational nodal code, VARIANT. We employ the GMRES algorithm to treat non-symmetric response matrix equations. A pre conditioner is formulated for the within-group diffusion equation which is equivalent to partitioned matrix acceleration of RBGS iterations. We employ the pre conditioner, which closely parallels two-level p multigrid, to improve RBGS and GMRES algorithms. Of the accelerated algorithms, GMRES converges with less computational effort than RBGS and therefore is chosen for further development. The p multigrid pre conditioner requires response matrices with two or more degrees of freedom (DOF) per interface that are polynomials, which are both orthogonal and hierarchical. It is therefore not directly applicable to very fine mesh calculations that are both slow to converge and that are often modeled with response matrices with only one DOF per interface. Orthogonal matrix aggregation (OMA) is introduced to circumvent this difficulty by combining N×N fine mesh response matrices with one DOF per interface into a coarse mesh response matrix with N orthogonal DOF per interface. Numerical results show that OMA used alone or in combination with p multigrid preconditioning substantially accelerates GMRES solutions. (author)

  4. Preconditioned Krylov and Gauss-Seidel solutions of response matrix equations

    Energy Technology Data Exchange (ETDEWEB)

    Lewis, E.E., E-mail: e-lewis@northwestern.edu [Department of Mechanical Engineering, Northwestern University, Evanston, IL (United States); Smith, M.A.; Yang, W.S.; Wollaber, A., E-mail: masmith@anl.gov, E-mail: wsyang@anl.gov, E-mail: wollaber@lanl.gov [Nuclear Engineering Division, Argonne National Laboratory, Argonne, IL (United States)

    2011-07-01

    The use of preconditioned Krylov methods is examined as an alternative to the partitioned matrix acceleration applied to red-black Gauss Seidel (RBGS) iteration that is presently used in the variational nodal code, VARIANT. We employ the GMRES algorithm to treat non-symmetric response matrix equations. A pre conditioner is formulated for the within-group diffusion equation which is equivalent to partitioned matrix acceleration of RBGS iterations. We employ the pre conditioner, which closely parallels two-level p multigrid, to improve RBGS and GMRES algorithms. Of the accelerated algorithms, GMRES converges with less computational effort than RBGS and therefore is chosen for further development. The p multigrid pre conditioner requires response matrices with two or more degrees of freedom (DOF) per interface that are polynomials, which are both orthogonal and hierarchical. It is therefore not directly applicable to very fine mesh calculations that are both slow to converge and that are often modeled with response matrices with only one DOF per interface. Orthogonal matrix aggregation (OMA) is introduced to circumvent this difficulty by combining N×N fine mesh response matrices with one DOF per interface into a coarse mesh response matrix with N orthogonal DOF per interface. Numerical results show that OMA used alone or in combination with p multigrid preconditioning substantially accelerates GMRES solutions. (author)

  5. Krylov solvers preconditioned with the low-order red-black algorithm for the PN hybrid FEM for the instant code

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Yaqi; Rabiti, Cristian; Palmiotti, Giuseppe, E-mail: yaqi.wang@inl.gov, E-mail: cristian.rabiti@inl.gov, E-mail: giuseppe.palmiotti@inl.gov [Idaho National Laboratory, Idaho Falls, ID (United States)

    2011-07-01

    This paper proposes a new set of Krylov solvers, CG and GMRes, as an alternative of the Red-Black (RB) algorithm on on solving the steady-state one-speed neutron transport equation discretized with PN in angle and hybrid FEM (Finite Element Method) in space. A pre conditioner with the low-order RB iteration is designed to improve their convergence. These Krylov solvers can reduce the cost of pre-assembling the response matrices greatly. Numerical results with the INSTANT code are presented in order to show that they can be a good supplement on solving the PN-HFEM system. (author)

  6. Krylov solvers preconditioned with the low-order red-black algorithm for the PN hybrid FEM for the instant code

    International Nuclear Information System (INIS)

    Wang, Yaqi; Rabiti, Cristian; Palmiotti, Giuseppe

    2011-01-01

    This paper proposes a new set of Krylov solvers, CG and GMRes, as an alternative of the Red-Black (RB) algorithm on on solving the steady-state one-speed neutron transport equation discretized with PN in angle and hybrid FEM (Finite Element Method) in space. A pre conditioner with the low-order RB iteration is designed to improve their convergence. These Krylov solvers can reduce the cost of pre-assembling the response matrices greatly. Numerical results with the INSTANT code are presented in order to show that they can be a good supplement on solving the PN-HFEM system. (author)

  7. An Image Segmentation Based on a Genetic Algorithm for Determining Soil Coverage by Crop Residues

    Science.gov (United States)

    Ribeiro, Angela; Ranz, Juan; Burgos-Artizzu, Xavier P.; Pajares, Gonzalo; Sanchez del Arco, Maria J.; Navarrete, Luis

    2011-01-01

    Determination of the soil coverage by crop residues after ploughing is a fundamental element of Conservation Agriculture. This paper presents the application of genetic algorithms employed during the fine tuning of the segmentation process of a digital image with the aim of automatically quantifying the residue coverage. In other words, the objective is to achieve a segmentation that would permit the discrimination of the texture of the residue so that the output of the segmentation process is a binary image in which residue zones are isolated from the rest. The RGB images used come from a sample of images in which sections of terrain were photographed with a conventional camera positioned in zenith orientation atop a tripod. The images were taken outdoors under uncontrolled lighting conditions. Up to 92% similarity was achieved between the images obtained by the segmentation process proposed in this paper and the templates made by an elaborate manual tracing process. In addition to the proposed segmentation procedure and the fine tuning procedure that was developed, a global quantification of the soil coverage by residues for the sampled area was achieved that differed by only 0.85% from the quantification obtained using template images. Moreover, the proposed method does not depend on the type of residue present in the image. The study was conducted at the experimental farm “El Encín” in Alcalá de Henares (Madrid, Spain). PMID:22163966

  8. Model parameter estimations from residual gravity anomalies due to simple-shaped sources using Differential Evolution Algorithm

    Science.gov (United States)

    Ekinci, Yunus Levent; Balkaya, Çağlayan; Göktürkler, Gökhan; Turan, Seçil

    2016-06-01

    An efficient approach to estimate model parameters from residual gravity data based on differential evolution (DE), a stochastic vector-based metaheuristic algorithm, has been presented. We have showed the applicability and effectiveness of this algorithm on both synthetic and field anomalies. According to our knowledge, this is a first attempt of applying DE for the parameter estimations of residual gravity anomalies due to isolated causative sources embedded in the subsurface. The model parameters dealt with here are the amplitude coefficient (A), the depth and exact origin of causative source (zo and xo, respectively) and the shape factors (q and ƞ). The error energy maps generated for some parameter pairs have successfully revealed the nature of the parameter estimation problem under consideration. Noise-free and noisy synthetic single gravity anomalies have been evaluated with success via DE/best/1/bin, which is a widely used strategy in DE. Additionally some complicated gravity anomalies caused by multiple source bodies have been considered, and the results obtained have showed the efficiency of the algorithm. Then using the strategy applied in synthetic examples some field anomalies observed for various mineral explorations such as a chromite deposit (Camaguey district, Cuba), a manganese deposit (Nagpur, India) and a base metal sulphide deposit (Quebec, Canada) have been considered to estimate the model parameters of the ore bodies. Applications have exhibited that the obtained results such as the depths and shapes of the ore bodies are quite consistent with those published in the literature. Uncertainty in the solutions obtained from DE algorithm has been also investigated by Metropolis-Hastings (M-H) sampling algorithm based on simulated annealing without cooling schedule. Based on the resulting histogram reconstructions of both synthetic and field data examples the algorithm has provided reliable parameter estimations being within the sampling limits of

  9. Development of numerical solution techniques in the KIKO3D code

    International Nuclear Information System (INIS)

    Panka, Istvan; Kereszturi, Andras; Hegedus, Csaba

    2005-01-01

    The paper describes the numerical methods applied in KIKO3D three-dimensional reactor dynamics code and present a new, more effective method (Bi-CGSTAB) for accelerating the large sparse matrix equation solution. The convergence characteristics were investigated in a given macro time step of a Control Rod Ejection transient. The results obtained by the old GMRES and new Bi-CGSTAB methods are compared. It is concluded that the real relative errors of the solutions obtained by GMRES or Bi - CGSTAB algorithms are in fact closer together than the estimated relative errors. The KIKO3D-Bi-CGSTAB method converges safely and it is 7-12 % faster than the old KIKO3D-GMRES solution (Authors)

  10. Classification of different kinds of pesticide residues on lettuce based on fluorescence spectra and WT-BCC-SVM algorithm

    Science.gov (United States)

    Zhou, Xin; Jun, Sun; Zhang, Bing; Jun, Wu

    2017-07-01

    In order to improve the reliability of the spectrum feature extracted by wavelet transform, a method combining wavelet transform (WT) with bacterial colony chemotaxis algorithm and support vector machine (BCC-SVM) algorithm (WT-BCC-SVM) was proposed in this paper. Besides, we aimed to identify different kinds of pesticide residues on lettuce leaves in a novel and rapid non-destructive way by using fluorescence spectra technology. The fluorescence spectral data of 150 lettuce leaf samples of five different kinds of pesticide residues on the surface of lettuce were obtained using Cary Eclipse fluorescence spectrometer. Standard normalized variable detrending (SNV detrending), Savitzky-Golay coupled with Standard normalized variable detrending (SG-SNV detrending) were used to preprocess the raw spectra, respectively. Bacterial colony chemotaxis combined with support vector machine (BCC-SVM) and support vector machine (SVM) classification models were established based on full spectra (FS) and wavelet transform characteristics (WTC), respectively. Moreover, WTC were selected by WT. The results showed that the accuracy of training set, calibration set and the prediction set of the best optimal classification model (SG-SNV detrending-WT-BCC-SVM) were 100%, 98% and 93.33%, respectively. In addition, the results indicated that it was feasible to use WT-BCC-SVM to establish diagnostic model of different kinds of pesticide residues on lettuce leaves.

  11. Numerical simulations of microwave heating of liquids: enhancements using Krylov subspace methods

    Science.gov (United States)

    Lollchund, M. R.; Dookhitram, K.; Sunhaloo, M. S.; Boojhawon, R.

    2013-04-01

    In this paper, we compare the performances of three iterative solvers for large sparse linear systems arising in the numerical computations of incompressible Navier-Stokes (NS) equations. These equations are employed mainly in the simulation of microwave heating of liquids. The emphasis of this work is on the application of Krylov projection techniques such as Generalized Minimal Residual (GMRES) to solve the Pressure Poisson Equations that result from discretisation of the NS equations. The performance of the GMRES method is compared with the traditional Gauss-Seidel (GS) and point successive over relaxation (PSOR) techniques through their application to simulate the dynamics of water housed inside a vertical cylindrical vessel which is subjected to microwave radiation. It is found that as the mesh size increases, GMRES gives the fastest convergence rate in terms of computational times and number of iterations.

  12. Numerical simulations of microwave heating of liquids: enhancements using Krylov subspace methods

    International Nuclear Information System (INIS)

    Lollchund, M R; Dookhitram, K; Sunhaloo, M S; Boojhawon, R

    2013-01-01

    In this paper, we compare the performances of three iterative solvers for large sparse linear systems arising in the numerical computations of incompressible Navier-Stokes (NS) equations. These equations are employed mainly in the simulation of microwave heating of liquids. The emphasis of this work is on the application of Krylov projection techniques such as Generalized Minimal Residual (GMRES) to solve the Pressure Poisson Equations that result from discretisation of the NS equations. The performance of the GMRES method is compared with the traditional Gauss-Seidel (GS) and point successive over relaxation (PSOR) techniques through their application to simulate the dynamics of water housed inside a vertical cylindrical vessel which is subjected to microwave radiation. It is found that as the mesh size increases, GMRES gives the fastest convergence rate in terms of computational times and number of iterations.

  13. An incomplete assembly with thresholding algorithm for systems of reaction-diffusion equations in three space dimensions IAT for reaction-diffusion systems

    International Nuclear Information System (INIS)

    Moore, Peter K.

    2003-01-01

    Solving systems of reaction-diffusion equations in three space dimensions can be prohibitively expensive both in terms of storage and CPU time. Herein, I present a new incomplete assembly procedure that is designed to reduce storage requirements. Incomplete assembly is analogous to incomplete factorization in that only a fixed number of nonzero entries are stored per row and a drop tolerance is used to discard small values. The algorithm is incorporated in a finite element method-of-lines code and tested on a set of reaction-diffusion systems. The effect of incomplete assembly on CPU time and storage and on the performance of the temporal integrator DASPK, algebraic solver GMRES and preconditioner ILUT is studied

  14. A Framework for Generalized Conjugate Gradient Methods - with Special Emphasis on Contributions by Rüdiger Weiss

    Czech Academy of Sciences Publication Activity Database

    Gutknecht, M. H.; Rozložník, Miroslav

    2002-01-01

    Roč. 41, - (2002), s. 7-22 ISSN 0168-9274 R&D Projects: GA AV ČR IAA1030103; GA ČR GA101/00/1035 Institutional research plan: AV0Z1030915 Keywords : sparse linear systems * Krylov space method * orthogonal residual method * minimal residual method * conjugate gradient method * residual smoothing * CG * CGNE * CGNR * CR * FOM * GMRES * PRES Subject RIV: BA - General Mathematics Impact factor: 0.504, year: 2002

  15. Aplicação do pré-condicionador Multigrid Algébrico baseado em Wavelet no cálculo de campos magnéticos não lineares

    Directory of Open Access Journals (Sweden)

    Fabio Henrique Pereira

    2009-01-01

    Full Text Available In this work the performance of ¿-cycle wavelet-based algebraic multigrid preconditioner for iterative methods is investigated. The method is applied as a preconditioner for the classical iterative methods Bi-Conjugate Gradient Stabilized (BiCGStab, Generalized Minimum Residual (GMRes and Conjugate Gradient (CG to the solution of non-linear system of algebraic equations from the analysis of a switched reluctance motor with ferromagnetic material the steel S45C and nonlinear magnetization curve, associated with the Newton-Raphson algorithm. Particular attention has been focused in both V- and W-cycle convergence factors, as well as the CPU time. Numerical results show the efficiency of the proposed techniques when compared with classical preconditioner, such as Incomplete Cholesky and Incomplete LU decomposition.

  16. Exact Solutions for Internuclear Vectors and Backbone Dihedral Angles from NH Residual Dipolar Couplings in Two Media, and their Application in a Systematic Search Algorithm for Determining Protein Backbone Structure

    International Nuclear Information System (INIS)

    Wang Lincong; Donald, Bruce Randall

    2004-01-01

    We have derived a quartic equation for computing the direction of an internuclear vector from residual dipolar couplings (RDCs) measured in two aligning media, and two simple trigonometric equations for computing the backbone (φ,ψ) angles from two backbone vectors in consecutive peptide planes. These equations make it possible to compute, exactly and in constant time, the backbone (φ,ψ) angles for a residue from RDCs in two media on any single backbone vector type. Building upon these exact solutions we have designed a novel algorithm for determining a protein backbone substructure consisting of α-helices and β-sheets. Our algorithm employs a systematic search technique to refine the conformation of both α-helices and β-sheets and to determine their orientations using exclusively the angular restraints from RDCs. The algorithm computes the backbone substructure employing very sparse distance restraints between pairs of α-helices and β-sheets refined by the systematic search. The algorithm has been demonstrated on the protein human ubiquitin using only backbone NH RDCs, plus twelve hydrogen bonds and four NOE distance restraints. Further, our results show that both the global orientations and the conformations of α-helices and β-strands can be determined with high accuracy using only two RDCs per residue. The algorithm requires, as its input, backbone resonance assignments, the identification of α-helices and β-sheets as well as sparse NOE distance and hydrogen bond restraints.Abbreviations: NMR - nuclear magnetic resonance; RDC - residual dipolar coupling; NOE - nuclear Overhauser effect; SVD - singular value decomposition; DFS - depth-first search; RMSD - root mean square deviation; POF - principal order frame; PDB - protein data bank; SA - simulated annealing; MD - molecular dynamics

  17. Efficient parallel implicit methods for rotary-wing aerodynamics calculations

    Science.gov (United States)

    Wissink, Andrew M.

    Euler/Navier-Stokes Computational Fluid Dynamics (CFD) methods are commonly used for prediction of the aerodynamics and aeroacoustics of modern rotary-wing aircraft. However, their widespread application to large complex problems is limited lack of adequate computing power. Parallel processing offers the potential for dramatic increases in computing power, but most conventional implicit solution methods are inefficient in parallel and new techniques must be adopted to realize its potential. This work proposes alternative implicit schemes for Euler/Navier-Stokes rotary-wing calculations which are robust and efficient in parallel. The first part of this work proposes an efficient parallelizable modification of the Lower Upper-Symmetric Gauss Seidel (LU-SGS) implicit operator used in the well-known Transonic Unsteady Rotor Navier Stokes (TURNS) code. The new hybrid LU-SGS scheme couples a point-relaxation approach of the Data Parallel-Lower Upper Relaxation (DP-LUR) algorithm for inter-processor communication with the Symmetric Gauss Seidel algorithm of LU-SGS for on-processor computations. With the modified operator, TURNS is implemented in parallel using Message Passing Interface (MPI) for communication. Numerical performance and parallel efficiency are evaluated on the IBM SP2 and Thinking Machines CM-5 multi-processors for a variety of steady-state and unsteady test cases. The hybrid LU-SGS scheme maintains the numerical performance of the original LU-SGS algorithm in all cases and shows a good degree of parallel efficiency. It experiences a higher degree of robustness than DP-LUR for third-order upwind solutions. The second part of this work examines use of Krylov subspace iterative solvers for the nonlinear CFD solutions. The hybrid LU-SGS scheme is used as a parallelizable preconditioner. Two iterative methods are tested, Generalized Minimum Residual (GMRES) and Orthogonal s-Step Generalized Conjugate Residual (OSGCR). The Newton method demonstrates good

  18. Research on wind field algorithm of wind lidar based on BP neural network and grey prediction

    Science.gov (United States)

    Chen, Yong; Chen, Chun-Li; Luo, Xiong; Zhang, Yan; Yang, Ze-hou; Zhou, Jie; Shi, Xiao-ding; Wang, Lei

    2018-01-01

    This paper uses the BP neural network and grey algorithm to forecast and study radar wind field. In order to reduce the residual error in the wind field prediction which uses BP neural network and grey algorithm, calculating the minimum value of residual error function, adopting the residuals of the gray algorithm trained by BP neural network, using the trained network model to forecast the residual sequence, using the predicted residual error sequence to modify the forecast sequence of the grey algorithm. The test data show that using the grey algorithm modified by BP neural network can effectively reduce the residual value and improve the prediction precision.

  19. Residue preference mapping of ligand fragments in the Protein Data Bank.

    Science.gov (United States)

    Wang, Lirong; Xie, Zhaojun; Wipf, Peter; Xie, Xiang-Qun

    2011-04-25

    The interaction between small molecules and proteins is one of the major concerns for structure-based drug design because the principles of protein-ligand interactions and molecular recognition are not thoroughly understood. Fortunately, the analysis of protein-ligand complexes in the Protein Data Bank (PDB) enables unprecedented possibilities for new insights. Herein, we applied molecule-fragmentation algorithms to split the ligands extracted from PDB crystal structures into small fragments. Subsequently, we have developed a ligand fragment and residue preference mapping (LigFrag-RPM) algorithm to map the profiles of the interactions between these fragments and the 20 proteinogenic amino acid residues. A total of 4032 fragments were generated from 71 798 PDB ligands by a ring cleavage (RC) algorithm. Among these ligand fragments, 315 unique fragments were characterized with the corresponding fragment-residue interaction profiles by counting residues close to these fragments. The interaction profiles revealed that these fragments have specific preferences for certain types of residues. The applications of these interaction profiles were also explored and evaluated in case studies, showing great potential for the study of protein-ligand interactions and drug design. Our studies demonstrated that the fragment-residue interaction profiles generated from the PDB ligand fragments can be used to detect whether these fragments are in their favorable or unfavorable environments. The algorithm for a ligand fragment and residue preference mapping (LigFrag-RPM) developed here also has the potential to guide lead chemistry modifications as well as binding residues predictions.

  20. Automatic residue removal for high-NA extreme illumination

    Science.gov (United States)

    Moon, James; Nam, Byong-Sub; Jeong, Joo-Hong; Kong, Dong-Ho; Nam, Byung-Ho; Yim, Dong Gyu

    2007-10-01

    An epidemic for smaller node has been that, as the device architecture shrinks, lithography process requires high Numerical Aperture (NA), and extreme illumination system. This, in turn, creates many lithography problems such as low lithography process margin (Depth of Focus, Exposure Latitude), unstable Critical Dimension (CD) uniformity and restricted guideline for device design rule and so on. Especially for high NA, extreme illumination such as immersion illumination systems, above all the related problems, restricted design rule due to forbidden pitch is critical and crucial issue. This forbidden pitch is composed of numerous optical effects but majority of these forbidden pitch compose of photo resist residue and these residue must be removed to relieve some room for already tight design rule. In this study, we propose automated algorithm to remove photo resist residue due to high NA and extreme illumination condition. This algorithm automatically self assembles assist patterns based on the original design layout, therefore insuring the safety and simplicity of the generated assist pattern to the original design and removes any resist residue created by extreme illumination condition. Also we tested our automated algorithm on full chip FLASH memory device and showed the residue removal effect by using commercial verification tools as well as on actual test wafer.

  1. Quantum Algorithms for Weighing Matrices and Quadratic Residues

    OpenAIRE

    van Dam, Wim

    2000-01-01

    In this article we investigate how we can employ the structure of combinatorial objects like Hadamard matrices and weighing matrices to device new quantum algorithms. We show how the properties of a weighing matrix can be used to construct a problem for which the quantum query complexity is ignificantly lower than the classical one. It is pointed out that this scheme captures both Bernstein & Vazirani's inner-product protocol, as well as Grover's search algorithm. In the second part of the ar...

  2. Adaptive Residual Interpolation for Color and Multispectral Image Demosaicking.

    Science.gov (United States)

    Monno, Yusuke; Kiku, Daisuke; Tanaka, Masayuki; Okutomi, Masatoshi

    2017-12-01

    Color image demosaicking for the Bayer color filter array is an essential image processing operation for acquiring high-quality color images. Recently, residual interpolation (RI)-based algorithms have demonstrated superior demosaicking performance over conventional color difference interpolation-based algorithms. In this paper, we propose adaptive residual interpolation (ARI) that improves existing RI-based algorithms by adaptively combining two RI-based algorithms and selecting a suitable iteration number at each pixel. These are performed based on a unified criterion that evaluates the validity of an RI-based algorithm. Experimental comparisons using standard color image datasets demonstrate that ARI can improve existing RI-based algorithms by more than 0.6 dB in the color peak signal-to-noise ratio and can outperform state-of-the-art algorithms based on training images. We further extend ARI for a multispectral filter array, in which more than three spectral bands are arrayed, and demonstrate that ARI can achieve state-of-the-art performance also for the task of multispectral image demosaicking.

  3. Identification of residue pairing in interacting β-strands from a predicted residue contact map.

    Science.gov (United States)

    Mao, Wenzhi; Wang, Tong; Zhang, Wenxuan; Gong, Haipeng

    2018-04-19

    Despite the rapid progress of protein residue contact prediction, predicted residue contact maps frequently contain many errors. However, information of residue pairing in β strands could be extracted from a noisy contact map, due to the presence of characteristic contact patterns in β-β interactions. This information may benefit the tertiary structure prediction of mainly β proteins. In this work, we propose a novel ridge-detection-based β-β contact predictor to identify residue pairing in β strands from any predicted residue contact map. Our algorithm RDb 2 C adopts ridge detection, a well-developed technique in computer image processing, to capture consecutive residue contacts, and then utilizes a novel multi-stage random forest framework to integrate the ridge information and additional features for prediction. Starting from the predicted contact map of CCMpred, RDb 2 C remarkably outperforms all state-of-the-art methods on two conventional test sets of β proteins (BetaSheet916 and BetaSheet1452), and achieves F1-scores of ~ 62% and ~ 76% at the residue level and strand level, respectively. Taking the prediction of the more advanced RaptorX-Contact as input, RDb 2 C achieves impressively higher performance, with F1-scores reaching ~ 76% and ~ 86% at the residue level and strand level, respectively. In a test of structural modeling using the top 1 L predicted contacts as constraints, for 61 mainly β proteins, the average TM-score achieves 0.442 when using the raw RaptorX-Contact prediction, but increases to 0.506 when using the improved prediction by RDb 2 C. Our method can significantly improve the prediction of β-β contacts from any predicted residue contact maps. Prediction results of our algorithm could be directly applied to effectively facilitate the practical structure prediction of mainly β proteins. All source data and codes are available at http://166.111.152.91/Downloads.html or the GitHub address of https://github.com/wzmao/RDb2C .

  4. A balancing domain decomposition method by constraints for advection-diffusion problems

    Energy Technology Data Exchange (ETDEWEB)

    Tu, Xuemin; Li, Jing

    2008-12-10

    The balancing domain decomposition methods by constraints are extended to solving nonsymmetric, positive definite linear systems resulting from the finite element discretization of advection-diffusion equations. A pre-conditioned GMRES iteration is used to solve a Schur complement system of equations for the subdomain interface variables. In the preconditioning step of each iteration, a partially sub-assembled finite element problem is solved. A convergence rate estimate for the GMRES iteration is established, under the condition that the diameters of subdomains are small enough. It is independent of the number of subdomains and grows only slowly with the subdomain problem size. Numerical experiments for several two-dimensional advection-diffusion problems illustrate the fast convergence of the proposed algorithm.

  5. A parallel additive Schwarz preconditioned Jacobi-Davidson algorithm for polynomial eigenvalue problems in quantum dot simulation

    International Nuclear Information System (INIS)

    Hwang, F-N; Wei, Z-H; Huang, T-M; Wang Weichung

    2010-01-01

    We develop a parallel Jacobi-Davidson approach for finding a partial set of eigenpairs of large sparse polynomial eigenvalue problems with application in quantum dot simulation. A Jacobi-Davidson eigenvalue solver is implemented based on the Portable, Extensible Toolkit for Scientific Computation (PETSc). The eigensolver thus inherits PETSc's efficient and various parallel operations, linear solvers, preconditioning schemes, and easy usages. The parallel eigenvalue solver is then used to solve higher degree polynomial eigenvalue problems arising in numerical simulations of three dimensional quantum dots governed by Schroedinger's equations. We find that the parallel restricted additive Schwarz preconditioner in conjunction with a parallel Krylov subspace method (e.g. GMRES) can solve the correction equations, the most costly step in the Jacobi-Davidson algorithm, very efficiently in parallel. Besides, the overall performance is quite satisfactory. We have observed near perfect superlinear speedup by using up to 320 processors. The parallel eigensolver can find all target interior eigenpairs of a quintic polynomial eigenvalue problem with more than 32 million variables within 12 minutes by using 272 Intel 3.0 GHz processors.

  6. On the comparison of perturbation-iteration algorithm and residual power series method to solve fractional Zakharov-Kuznetsov equation

    Science.gov (United States)

    Şenol, Mehmet; Alquran, Marwan; Kasmaei, Hamed Daei

    2018-06-01

    In this paper, we present analytic-approximate solution of time-fractional Zakharov-Kuznetsov equation. This model demonstrates the behavior of weakly nonlinear ion acoustic waves in a plasma bearing cold ions and hot isothermal electrons in the presence of a uniform magnetic field. Basic definitions of fractional derivatives are described in the Caputo sense. Perturbation-iteration algorithm (PIA) and residual power series method (RPSM) are applied to solve this equation with success. The convergence analysis is also presented for both methods. Numerical results are given and then they are compared with the exact solutions. Comparison of the results reveal that both methods are competitive, powerful, reliable, simple to use and ready to apply to wide range of fractional partial differential equations.

  7. Prediction of interface residue based on the features of residue interaction network.

    Science.gov (United States)

    Jiao, Xiong; Ranganathan, Shoba

    2017-11-07

    Protein-protein interaction plays a crucial role in the cellular biological processes. Interface prediction can improve our understanding of the molecular mechanisms of the related processes and functions. In this work, we propose a classification method to recognize the interface residue based on the features of a weighted residue interaction network. The random forest algorithm is used for the prediction and 16 network parameters and the B-factor are acting as the element of the input feature vector. Compared with other similar work, the method is feasible and effective. The relative importance of these features also be analyzed to identify the key feature for the prediction. Some biological meaning of the important feature is explained. The results of this work can be used for the related work about the structure-function relationship analysis via a residue interaction network model. Copyright © 2017 Elsevier Ltd. All rights reserved.

  8. A Residual Approach for Balanced Truncation Model Reduction (BTMR of Compartmental Systems

    Directory of Open Access Journals (Sweden)

    William La Cruz

    2014-05-01

    Full Text Available This paper presents a residual approach of the square root balanced truncation algorithm for model order reduction of continuous, linear and time-invariante compartmental systems. Specifically, the new approach uses a residual method to approximate the controllability and observability gramians, whose resolution is an essential step of the square root balanced truncation algorithm, that requires a great computational cost. Numerical experiences are included to highlight the efficacy of the proposed approach.

  9. VCODE, Ordinary Differential Equation Solver for Stiff and Non-Stiff Problems

    International Nuclear Information System (INIS)

    Cohen, Scott D.; Hindmarsh, Alan C.

    2001-01-01

    1 - Description of program or function: CVODE is a package written in ANSI standard C for solving initial value problems for ordinary differential equations. It solves both stiff and non stiff systems. In the stiff case, it includes a variety of options for treating the Jacobian of the system, including dense and band matrix solvers, and a preconditioned Krylov (iterative) solver. 2 - Method of solution: Integration is by Adams or BDF (Backward Differentiation Formula) methods, at user option. Corrector iteration is by functional iteration or Newton iteration. For the solution of linear systems within Newton iteration, users can select a dense solver, a band solver, a diagonal approximation, or a preconditioned Generalized Minimal Residual (GMRES) solver. In the dense and band cases, the user can supply a Jacobian approximation or let CVODE generate it internally. In the GMRES case, the pre-conditioner is user-supplied

  10. Simulation of 3D parachute fluid–structure interaction based on nonlinear finite element method and preconditioning finite volume method

    Directory of Open Access Journals (Sweden)

    Fan Yuxin

    2014-12-01

    Full Text Available A fluid–structure interaction method combining a nonlinear finite element algorithm with a preconditioning finite volume method is proposed in this paper to simulate parachute transient dynamics. This method uses a three-dimensional membrane–cable fabric model to represent a parachute system at a highly folded configuration. The large shape change during parachute inflation is computed by the nonlinear Newton–Raphson iteration and the linear system equation is solved by the generalized minimal residual (GMRES method. A membrane wrinkling algorithm is also utilized to evaluate the special uniaxial tension state of membrane elements on the parachute canopy. In order to avoid large time expenses during structural nonlinear iteration, the implicit Hilber–Hughes–Taylor (HHT time integration method is employed. For the fluid dynamic simulations, the Roe and HLLC (Harten–Lax–van Leer contact scheme has been modified and extended to compute flow problems at all speeds. The lower–upper symmetric Gauss–Seidel (LU-SGS approximate factorization is applied to accelerate the numerical convergence speed. Finally, the test model of a highly folded C-9 parachute is simulated at a prescribed speed and the results show similar characteristics compared with experimental results and previous literature.

  11. Model-based leakage localization in drinking water distribution networks using structured residuals

    OpenAIRE

    Puig Cayuela, Vicenç; Rosich, Albert

    2013-01-01

    In this paper, a new model based approach to leakage localization in drinking water networks is proposed based on generating a set of structured residuals. The residual evaluation is based on a numerical method based on an enhanced Newton-Raphson algorithm. The proposed method is suitable for water network systems because the non-linearities of the model make impossible to derive analytical residuals. Furthermore, the computed residuals are designed so that leaks are decoupled, which impro...

  12. Deploy Nalu/Kokkos algorithmic infrastructure with performance benchmarking.

    Energy Technology Data Exchange (ETDEWEB)

    Domino, Stefan P. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Ananthan, Shreyas [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Knaus, Robert C. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Williams, Alan B. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2017-09-29

    The former Nalu interior heterogeneous algorithm design, which was originally designed to manage matrix assembly operations over all elemental topology types, has been modified to operate over homogeneous collections of mesh entities. This newly templated kernel design allows for removal of workset variable resize operations that were formerly required at each loop over a Sierra ToolKit (STK) bucket (nominally, 512 entities in size). Extensive usage of the Standard Template Library (STL) std::vector has been removed in favor of intrinsic Kokkos memory views. In this milestone effort, the transition to Kokkos as the underlying infrastructure to support performance and portability on many-core architectures has been deployed for key matrix algorithmic kernels. A unit-test driven design effort has developed a homogeneous entity algorithm that employs a team-based thread parallelism construct. The STK Single Instruction Multiple Data (SIMD) infrastructure is used to interleave data for improved vectorization. The collective algorithm design, which allows for concurrent threading and SIMD management, has been deployed for the core low-Mach element- based algorithm. Several tests to ascertain SIMD performance on Intel KNL and Haswell architectures have been carried out. The performance test matrix includes evaluation of both low- and higher-order methods. The higher-order low-Mach methodology builds on polynomial promotion of the core low-order control volume nite element method (CVFEM). Performance testing of the Kokkos-view/SIMD design indicates low-order matrix assembly kernel speed-up ranging between two and four times depending on mesh loading and node count. Better speedups are observed for higher-order meshes (currently only P=2 has been tested) especially on KNL. The increased workload per element on higher-order meshes bene ts from the wide SIMD width on KNL machines. Combining multiple threads with SIMD on KNL achieves a 4.6x speedup over the baseline, with

  13. An acceleration technique for 2D MOC based on Krylov subspace and domain decomposition methods

    International Nuclear Information System (INIS)

    Zhang Hongbo; Wu Hongchun; Cao Liangzhi

    2011-01-01

    Highlights: → We convert MOC into linear system solved by GMRES as an acceleration method. → We use domain decomposition method to overcome the inefficiency on large matrices. → Parallel technology is applied and a matched ray tracing system is developed. → Results show good efficiency even in large-scale and strong scattering problems. → The emphasis is that the technique is geometry-flexible. - Abstract: The method of characteristics (MOC) has great geometrical flexibility but poor computational efficiency in neutron transport calculations. The generalized minimal residual (GMRES) method, a type of Krylov subspace method, is utilized to accelerate a 2D generalized geometry characteristics solver AutoMOC. In this technique, a form of linear algebraic equation system for angular flux moments and boundary fluxes is derived to replace the conventional characteristics sweep (i.e. inner iteration) scheme, and then the GMRES method is implemented as an efficient linear system solver. This acceleration method is proved to be reliable in theory and simple for implementation. Furthermore, as introducing no restriction in geometry treatment, it is suitable for acceleration of an arbitrary geometry MOC solver. However, it is observed that the speedup decreases when the matrix becomes larger. The spatial domain decomposition method and multiprocessing parallel technology are then employed to overcome the problem. The calculation domain is partitioned into several sub-domains. For each of them, a smaller matrix is established and solved by GMRES; and the adjacent sub-domains are coupled by 'inner-edges', where the trajectory mismatches are considered adequately. Moreover, a matched ray tracing system is developed on the basis of AutoCAD, which allows a user to define the sub-domains on demand conveniently. Numerical results demonstrate that the acceleration techniques are efficient without loss of accuracy, even in the case of large-scale and strong scattering

  14. Exploiting residual information in the parameter choice for discrete ill-posed problems

    DEFF Research Database (Denmark)

    Hansen, Per Christian; Kilmer, Misha E.; Kjeldsen, Rikke Høj

    2006-01-01

    Most algorithms for choosing the regularization parameter in a discrete ill-posed problem are based on the norm of the residual vector. In this work we propose a different approach, where we seek to use all the information available in the residual vector. We present important relations between...

  15. Using In Silico Fragmentation to Improve Routine Residue Screening in Complex Matrices

    Science.gov (United States)

    Kaufmann, Anton; Butcher, Patrick; Maden, Kathryn; Walker, Stephan; Widmer, Mirjam

    2017-12-01

    Targeted residue screening requires the use of reference substances in order to identify potential residues. This becomes a difficult issue when using multi-residue methods capable of analyzing several hundreds of analytes. Therefore, the capability of in silico fragmentation based on a structure database ("suspect screening") instead of physical reference substances for routine targeted residue screening was investigated. The detection of fragment ions that can be predicted or explained by in silico software was utilized to reduce the number of false positives. These "proof of principle" experiments were done with a tool that is integrated into a commercial MS vendor instrument operating software (UNIFI) as well as with a platform-independent MS tool (Mass Frontier). A total of 97 analytes belonging to different chemical families were separated by reversed phase liquid chromatography and detected in a data-independent acquisition (DIA) mode using ion mobility hyphenated with quadrupole time of flight mass spectrometry. The instrument was operated in the MSE mode with alternating low and high energy traces. The fragments observed from product ion spectra were investigated using a "chopping" bond disconnection algorithm and a rule-based algorithm. The bond disconnection algorithm clearly explained more analyte product ions and a greater percentage of the spectral abundance than the rule-based software (92 out of the 97 compounds produced ≥1 explainable fragment ions). On the other hand, tests with a complex blank matrix (bovine liver extract) indicated that the chopping algorithm reports significantly more false positive fragments than the rule based software. [Figure not available: see fulltext.

  16. A Distributed and Energy-Efficient Algorithm for Event K-Coverage in Underwater Sensor Networks

    Directory of Open Access Journals (Sweden)

    Peng Jiang

    2017-01-01

    Full Text Available For event dynamic K-coverage algorithms, each management node selects its assistant node by using a greedy algorithm without considering the residual energy and situations in which a node is selected by several events. This approach affects network energy consumption and balance. Therefore, this study proposes a distributed and energy-efficient event K-coverage algorithm (DEEKA. After the network achieves 1-coverage, the nodes that detect the same event compete for the event management node with the number of candidate nodes and the average residual energy, as well as the distance to the event. Second, each management node estimates the probability of its neighbor nodes’ being selected by the event it manages with the distance level, the residual energy level, and the number of dynamic coverage event of these nodes. Third, each management node establishes an optimization model that uses expectation energy consumption and the residual energy variance of its neighbor nodes and detects the performance of the events it manages as targets. Finally, each management node uses a constrained non-dominated sorting genetic algorithm (NSGA-II to obtain the Pareto set of the model and the best strategy via technique for order preference by similarity to an ideal solution (TOPSIS. The algorithm first considers the effect of harsh underwater environments on information collection and transmission. It also considers the residual energy of a node and a situation in which the node is selected by several other events. Simulation results show that, unlike the on-demand variable sensing K-coverage algorithm, DEEKA balances and reduces network energy consumption, thereby prolonging the network’s best service quality and lifetime.

  17. 2D-RBUC for efficient parallel compression of residuals

    Science.gov (United States)

    Đurđević, Đorđe M.; Tartalja, Igor I.

    2018-02-01

    In this paper, we present a method for lossless compression of residuals with an efficient SIMD parallel decompression. The residuals originate from lossy or near lossless compression of height fields, which are commonly used to represent models of terrains. The algorithm is founded on the existing RBUC method for compression of non-uniform data sources. We have adapted the method to capture 2D spatial locality of height fields, and developed the data decompression algorithm for modern GPU architectures already present even in home computers. In combination with the point-level SIMD-parallel lossless/lossy high field compression method HFPaC, characterized by fast progressive decompression and seamlessly reconstructed surface, the newly proposed method trades off small efficiency degradation for a non negligible compression ratio (measured up to 91%) benefit.

  18. A Data Forward Stepwise Fitting Algorithm Based on Orthogonal Function System

    Directory of Open Access Journals (Sweden)

    Li Han-Ju

    2017-01-01

    Full Text Available Data fitting is the main method of functional data analysis, and it is widely used in the fields of economy, social science, engineering technology and so on. Least square method is the main method of data fitting, but the least square method is not convergent, no memory property, big fitting error and it is easy to over fitting. Based on the orthogonal trigonometric function system, this paper presents a data forward stepwise fitting algorithm. This algorithm takes forward stepwise fitting strategy, each time using the nearest base function to fit the residual error generated by the previous base function fitting, which makes the residual mean square error minimum. In this paper, we theoretically prove the convergence, the memory property and the fitting error diminishing character for the algorithm. Experimental results show that the proposed algorithm is effective, and the fitting performance is better than that of the least square method and the forward stepwise fitting algorithm based on the non-orthogonal function system.

  19. Development of a General Modelling Methodology for Vacuum Residue Hydroconversion

    Directory of Open Access Journals (Sweden)

    Pereira de Oliveira L.

    2013-11-01

    Full Text Available This work concerns the development of a methodology for kinetic modelling of refining processes, and more specifically for vacuum residue conversion. The proposed approach allows to overcome the lack of molecular detail of the petroleum fractions and to simulate the transformation of the feedstock molecules into effluent molecules by means of a two-step procedure. In the first step, a synthetic mixture of molecules representing the feedstock for the process is generated via a molecular reconstruction method, termed SR-REM molecular reconstruction. In the second step, a kinetic Monte-Carlo method (kMC is used to simulate the conversion reactions on this mixture of molecules. The molecular reconstruction was applied to several petroleum residues and is illustrated for an Athabasca (Canada vacuum residue. The kinetic Monte-Carlo method is then described in detail. In order to validate this stochastic approach, a lumped deterministic model for vacuum residue conversion was simulated using Gillespie’s Stochastic Simulation Algorithm. Despite the fact that both approaches are based on very different hypotheses, the stochastic simulation algorithm simulates the conversion reactions with the same accuracy as the deterministic approach. The full-scale stochastic simulation approach using molecular-level reaction pathways provides high amounts of detail on the effluent composition and is briefly illustrated for Athabasca VR hydrocracking.

  20. A method for the estimation of the residual error in the SALP approach for fault tree analysis

    International Nuclear Information System (INIS)

    Astolfi, M.; Contini, S.

    1980-01-01

    The aim of this report is the illustration of the algorithms implemented in the SALP-MP code for the estimation of the residual error. These algorithms are of more general use, and it would be possible to implement them on all codes of the series SALP previously developed, as well as, with minor modifications, to analysis procedures based on 'top-down' approaches. At the time, combined 'top-down' - 'bottom up' procedures are being studied in order to take advantage from both approaches for further reduction of computer time and better estimation of the residual error, for which the developed algorithms are still applicable

  1. Preconditioned iterative methods for space-time fractional advection-diffusion equations

    Science.gov (United States)

    Zhao, Zhi; Jin, Xiao-Qing; Lin, Matthew M.

    2016-08-01

    In this paper, we propose practical numerical methods for solving a class of initial-boundary value problems of space-time fractional advection-diffusion equations. First, we propose an implicit method based on two-sided Grünwald formulae and discuss its stability and consistency. Then, we develop the preconditioned generalized minimal residual (preconditioned GMRES) method and preconditioned conjugate gradient normal residual (preconditioned CGNR) method with easily constructed preconditioners. Importantly, because resulting systems are Toeplitz-like, fast Fourier transform can be applied to significantly reduce the computational cost. We perform numerical experiments to demonstrate the efficiency of our preconditioners, even in cases with variable coefficients.

  2. A Review of Element-Based Galerkin Methods for Numerical Weather Prediction

    Science.gov (United States)

    2015-04-01

    weighted residuals because a linear system of algebraic equations in the unknowns q̂ is built by imposing that∫ Ω wRdΩ = 0, (30) where R = L−S is the (non...space; this spectral transform is evaluated using a combination of Fourier and 10 Simone Marras1 et al. Legendre transforms. We perform an elementary ...For the first algebraic system, the GMRES method with a simple diagonal preconditioning is efficient in most of the cases, and few iterations are

  3. Effect of Inspection Policies and Residual Value of Collected Used Products: A Mathematical Model and Genetic Algorithm for a Closed-Loop Green Manufacturing System

    Directory of Open Access Journals (Sweden)

    Byung Duk Song

    2017-09-01

    Full Text Available In the green manufacturing system that pursues the reuse of used products, the residual value of collected used products (CUP hugely affects a variety of managerial decisions to construct profitable and environmental remanufacturing plans. This paper deals with a closed-loop green manufacturing system for companies which perform both manufacturing with raw materials and remanufacturing with collected used products (CUP. The amount of CUP is assumed as a function of buy-back cost while the quality level of CUP, which means the residual value, follows a known distribution. In addition, the remanufacturing cost can differ according to the quality of the CUP. Moreover, nowadays companies are subject to existing environment-related laws such as Extended Producer Responsibility (EPR. Therefore, a company should collect more used products than its obligatory take-back quota or face fines from the government for not meeting its quota. Through the development of mathematical models, two kinds of inspection policies are examined to validate the efficiency of two different operation processes. To find a managerial solution, a genetic algorithm is proposed and tested with numerical examples.

  4. Minimal residual cone-beam reconstruction with attenuation correction in SPECT

    International Nuclear Information System (INIS)

    La, Valerie; Grangeat, Pierre

    1998-01-01

    This paper presents an iterative method based on the minimal residual algorithm for tomographic attenuation compensated reconstruction from attenuated cone-beam projections given the attenuation distribution. Unlike conjugate-gradient based reconstruction techniques, the proposed minimal residual based algorithm solves directly a quasisymmetric linear system, which is a preconditioned system. Thus it avoids the use of normal equations, which improves the convergence rate. Two main contributions are introduced. First, a regularization method is derived for quasisymmetric problems, based on a Tikhonov-Phillips regularization applied to the factorization of the symmetric part of the system matrix. This regularization is made spatially adaptive to avoid smoothing the region of interest. Second, our existing reconstruction algorithm for attenuation correction in parallel-beam geometry is extended to cone-beam geometry. A circular orbit is considered. Two preconditioning operators are proposed: the first one is Grangeat's inversion formula and the second one is Feldkamp's inversion formula. Experimental results obtained on simulated data are presented and the shadow zone effect on attenuated data is illustrated. (author)

  5. Peaks, plateaus, numerical instabilities, and achievable accuracy in Galerkin and norm minimizing procedures for solving Ax=b

    Energy Technology Data Exchange (ETDEWEB)

    Cullum, J. [IBM T.J. Watson Research Center, Yorktown Heights, NY (United States)

    1994-12-31

    Plots of the residual norms generated by Galerkin procedures for solving Ax = b often exhibit strings of irregular peaks. At seemingly erratic stages in the iterations, peaks appear in the residual norm plot, intervals of iterations over which the norms initially increase and then decrease. Plots of the residual norms generated by related norm minimizing procedures often exhibit long plateaus, sequences of iterations over which reductions in the size of the residual norm are unacceptably small. In an earlier paper the author discussed and derived relationships between such peaks and plateaus within corresponding Galerkin/Norm Minimizing pairs of such methods. In this paper, through a set of numerical experiments, the author examines connections between peaks, plateaus, numerical instabilities, and the achievable accuracy for such pairs of iterative methods. Three pairs of methods, GMRES/Arnoldi, QMR/BCG, and two bidiagonalization methods are studied.

  6. An improved single sensor parity space algorithm for sequential probability ratio test

    Energy Technology Data Exchange (ETDEWEB)

    Racz, A. [Hungarian Academy of Sciences, Budapest (Hungary). Atomic Energy Research Inst.

    1995-12-01

    In our paper we propose a modification of the single sensor parity algorithm in order to make the statistical properties of the generated residual determinable in advance. The algorithm is tested via computer simulated ramp failure at the temperature readings of the pressurizer. (author).

  7. Preconditioned conjugate gradient methods for the Navier-Stokes equations

    Science.gov (United States)

    Ajmani, Kumud; Ng, Wing-Fai; Liou, Meng-Sing

    1994-01-01

    A preconditioned Krylov subspace method (GMRES) is used to solve the linear systems of equations formed at each time-integration step of the unsteady, two-dimensional, compressible Navier-Stokes equations of fluid flow. The Navier-Stokes equations are cast in an implicit, upwind finite-volume, flux-split formulation. Several preconditioning techniques are investigated to enhance the efficiency and convergence rate of the implicit solver based on the GMRES algorithm. The superiority of the new solver is established by comparisons with a conventional implicit solver, namely line Gauss-Seidel relaxation (LGSR). Computational test results for low-speed (incompressible flow over a backward-facing step at Mach 0.1), transonic flow (trailing edge flow in a transonic turbine cascade), and hypersonic flow (shock-on-shock interactions on a cylindrical leading edge at Mach 6.0) are presented. For the Mach 0.1 case, overall speedup factors of up to 17 (in terms of time-steps) and 15 (in terms of CPU time on a CRAY-YMP/8) are found in favor of the preconditioned GMRES solver, when compared with the LGSR solver. The corresponding speedup factors for the transonic flow case are 17 and 23, respectively. The hypersonic flow case shows slightly lower speedup factors of 9 and 13, respectively. The study of preconditioners conducted in this research reveals that a new LUSGS-type preconditioner is much more efficient than a conventional incomplete LU-type preconditioner.

  8. MODEL FOR THE CORRECTION OF THE SPECIFIC GRAVITY OF BIODIESEL FROM RESIDUAL OIL

    Directory of Open Access Journals (Sweden)

    Tatiana Aparecida Rosa da Silva

    2013-06-01

    Full Text Available Biodiesel is a important fuel with economic benefits, social and environmental. The production cost of the biodiesel can be significantly lowered if the raw material is replaced by a alternative material as residual oil. In this study, the variation of specific gravity with temperature increase for diesel and biodiesel from residual oil obtained by homogeneous basic catalysis. All properties analyzed for biodiesel are within specification Brazil. The determination of the correction algorithm for the specific gravity function of temperature is also presented, and the slope of the line to diesel fuel, methylic biodiesel (BMR and ethylic biodiesel (BER from residual oil were respectively the values -0.7089, -0.7290 and -0.7277. This demonstrates the existence of difference of the model when compared chemically different fuels, like diesel and biodiesel from different sources, indicating the importance of determining the specific algorithm for the operations of conversion of volume to the reference temperature.

  9. The MARS15-based FermiCORD code system for calculation of the accelerator-induced residual dose

    Energy Technology Data Exchange (ETDEWEB)

    Grebe, A.; Leveling, A.; Lu, T.; Mokhov, N.; Pronskikh, V.

    2018-01-01

    The FermiCORD code system, a set of codes based on MARS15 that calculates the accelerator-induced residual doses at experimental facilities of arbitrary configurations, has been developed. FermiCORD is written in C++ as an add-on to Fortran-based MARS15. The FermiCORD algorithm consists of two stages: 1) simulation of residual doses on contact with the surfaces surrounding the studied location and of radionuclide inventories in the structures surrounding those locations using MARS15, and 2) simulation of the emission of the nuclear decay gamma-quanta by the residuals in the activated structures and scoring the prompt doses of these gamma-quanta at arbitrary distances from those structures. The FermiCORD code system has been benchmarked against similar algorithms based on other code systems and showed a good agreement. The code system has been applied for calculation of the residual dose of the target station for the Mu2e experiment and the results have been compared to approximate dosimetric approaches.

  10. HemeBIND: a novel method for heme binding residue prediction by combining structural and sequence information

    Directory of Open Access Journals (Sweden)

    Hu Jianjun

    2011-05-01

    Full Text Available Abstract Background Accurate prediction of binding residues involved in the interactions between proteins and small ligands is one of the major challenges in structural bioinformatics. Heme is an essential and commonly used ligand that plays critical roles in electron transfer, catalysis, signal transduction and gene expression. Although much effort has been devoted to the development of various generic algorithms for ligand binding site prediction over the last decade, no algorithm has been specifically designed to complement experimental techniques for identification of heme binding residues. Consequently, an urgent need is to develop a computational method for recognizing these important residues. Results Here we introduced an efficient algorithm HemeBIND for predicting heme binding residues by integrating structural and sequence information. We systematically investigated the characteristics of binding interfaces based on a non-redundant dataset of heme-protein complexes. It was found that several sequence and structural attributes such as evolutionary conservation, solvent accessibility, depth and protrusion clearly illustrate the differences between heme binding and non-binding residues. These features can then be separately used or combined to build the structure-based classifiers using support vector machine (SVM. The results showed that the information contained in these features is largely complementary and their combination achieved the best performance. To further improve the performance, an attempt has been made to develop a post-processing procedure to reduce the number of false positives. In addition, we built a sequence-based classifier based on SVM and sequence profile as an alternative when only sequence information can be used. Finally, we employed a voting method to combine the outputs of structure-based and sequence-based classifiers, which demonstrated remarkably better performance than the individual classifier alone

  11. A Method of Sky Ripple Residual Nonuniformity Reduction for a Cooled Infrared Imager and Hardware Implementation.

    Science.gov (United States)

    Li, Yiyang; Jin, Weiqi; Li, Shuo; Zhang, Xu; Zhu, Jin

    2017-05-08

    Cooled infrared detector arrays always suffer from undesired ripple residual nonuniformity (RNU) in sky scene observations. The ripple residual nonuniformity seriously affects the imaging quality, especially for small target detection. It is difficult to eliminate it using the calibration-based techniques and the current scene-based nonuniformity algorithms. In this paper, we present a modified temporal high-pass nonuniformity correction algorithm using fuzzy scene classification. The fuzzy scene classification is designed to control the correction threshold so that the algorithm can remove ripple RNU without degrading the scene details. We test the algorithm on a real infrared sequence by comparing it to several well-established methods. The result shows that the algorithm has obvious advantages compared with the tested methods in terms of detail conservation and convergence speed for ripple RNU correction. Furthermore, we display our architecture with a prototype built on a Xilinx Virtex-5 XC5VLX50T field-programmable gate array (FPGA), which has two advantages: (1) low resources consumption; and (2) small hardware delay (less than 10 image rows). It has been successfully applied in an actual system.

  12. 4D-PET reconstruction using a spline-residue model with spatial and temporal roughness penalties

    Science.gov (United States)

    Ralli, George P.; Chappell, Michael A.; McGowan, Daniel R.; Sharma, Ricky A.; Higgins, Geoff S.; Fenwick, John D.

    2018-05-01

    4D reconstruction of dynamic positron emission tomography (dPET) data can improve the signal-to-noise ratio in reconstructed image sequences by fitting smooth temporal functions to the voxel time-activity-curves (TACs) during the reconstruction, though the optimal choice of function remains an open question. We propose a spline-residue model, which describes TACs as weighted sums of convolutions of the arterial input function with cubic B-spline basis functions. Convolution with the input function constrains the spline-residue model at early time-points, potentially enhancing noise suppression in early time-frames, while still allowing a wide range of TAC descriptions over the entire imaged time-course, thus limiting bias. Spline-residue based 4D-reconstruction is compared to that of a conventional (non-4D) maximum a posteriori (MAP) algorithm, and to 4D-reconstructions based on adaptive-knot cubic B-splines, the spectral model and an irreversible two-tissue compartment (‘2C3K’) model. 4D reconstructions were carried out using a nested-MAP algorithm including spatial and temporal roughness penalties. The algorithms were tested using Monte-Carlo simulated scanner data, generated for a digital thoracic phantom with uptake kinetics based on a dynamic [18F]-Fluromisonidazole scan of a non-small cell lung cancer patient. For every algorithm, parametric maps were calculated by fitting each voxel TAC within a sub-region of the reconstructed images with the 2C3K model. Compared to conventional MAP reconstruction, spline-residue-based 4D reconstruction achieved  >50% improvements for five of the eight combinations of the four kinetics parameters for which parametric maps were created with the bias and noise measures used to analyse them, and produced better results for 5/8 combinations than any of the other reconstruction algorithms studied, while spectral model-based 4D reconstruction produced the best results for 2/8. 2C3K model-based 4D reconstruction generated

  13. Pattern-set generation algorithm for the one-dimensional multiple stock sizes cutting stock problem

    Science.gov (United States)

    Cui, Yaodong; Cui, Yi-Ping; Zhao, Zhigang

    2015-09-01

    A pattern-set generation algorithm (PSG) for the one-dimensional multiple stock sizes cutting stock problem (1DMSSCSP) is presented. The solution process contains two stages. In the first stage, the PSG solves the residual problems repeatedly to generate the patterns in the pattern set, where each residual problem is solved by the column-generation approach, and each pattern is generated by solving a single large object placement problem. In the second stage, the integer linear programming model of the 1DMSSCSP is solved using a commercial solver, where only the patterns in the pattern set are considered. The computational results of benchmark instances indicate that the PSG outperforms existing heuristic algorithms and rivals the exact algorithm in solution quality.

  14. Analysis of Pathfinder SST algorithm for global and regional conditions

    Indian Academy of Sciences (India)

    SST algorithms to improve the present accuracy of surface temperature measurements ... regions, except in the North Atlantic and adjacent seas, where the residuals are always positive. ..... the stratosphere causing significant contamination of.

  15. Reducing residual stresses and deformations in selective laser melting through multi-level multi-scale optimization of cellular scanning strategy

    DEFF Research Database (Denmark)

    Mohanty, Sankhya; Hattel, Jesper Henri

    2016-01-01

    . A multilevel optimization strategy is adopted using a customized genetic algorithm developed for optimizing cellular scanning strategy for selective laser melting, with an objective of reducing residual stresses and deformations. The resulting thermo-mechanically optimized cellular scanning strategies......, a calibrated, fast, multiscale thermal model coupled with a 3D finite element mechanical model is used to simulate residual stress formation and deformations during selective laser melting. The resulting reduction in thermal model computation time allows evolutionary algorithm-based optimization of the process...

  16. Fourier analysis of parallel block-Jacobi splitting with transport synthetic acceleration in two-dimensional geometry

    International Nuclear Information System (INIS)

    Rosa, M.; Warsa, J. S.; Chang, J. H.

    2007-01-01

    A Fourier analysis is conducted in two-dimensional (2D) Cartesian geometry for the discrete-ordinates (SN) approximation of the neutron transport problem solved with Richardson iteration (Source Iteration) and Richardson iteration preconditioned with Transport Synthetic Acceleration (TSA), using the Parallel Block-Jacobi (PBJ) algorithm. The results for the un-accelerated algorithm show that convergence of PBJ can degrade, leading in particular to stagnation of GMRES(m) in problems containing optically thin sub-domains. The results for the accelerated algorithm indicate that TSA can be used to efficiently precondition an iterative method in the optically thin case when implemented in the 'modified' version MTSA, in which only the scattering in the low order equations is reduced by some non-negative factor β<1. (authors)

  17. Predicting the concentration of residual methanol in industrial formalin using machine learning

    OpenAIRE

    Heidkamp, William

    2016-01-01

    In this thesis, a machine learning approach was used to develop a predictive model for residual methanol concentration in industrial formalin produced at the Akzo Nobel factory in Kristinehamn, Sweden. The MATLABTM computational environment supplemented with the Statistics and Machine LearningTM toolbox from the MathWorks were used to test various machine learning algorithms on the formalin production data from Akzo Nobel. As a result, the Gaussian Process Regression algorithm was found to pr...

  18. Efficient Dual Domain Decoding of Linear Block Codes Using Genetic Algorithms

    Directory of Open Access Journals (Sweden)

    Ahmed Azouaoui

    2012-01-01

    Full Text Available A computationally efficient algorithm for decoding block codes is developed using a genetic algorithm (GA. The proposed algorithm uses the dual code in contrast to the existing genetic decoders in the literature that use the code itself. Hence, this new approach reduces the complexity of decoding the codes of high rates. We simulated our algorithm in various transmission channels. The performance of this algorithm is investigated and compared with competitor decoding algorithms including Maini and Shakeel ones. The results show that the proposed algorithm gives large gains over the Chase-2 decoding algorithm and reach the performance of the OSD-3 for some quadratic residue (QR codes. Further, we define a new crossover operator that exploits the domain specific information and compare it with uniform and two point crossover. The complexity of this algorithm is also discussed and compared to other algorithms.

  19. Optimal Path Choice in Railway Passenger Travel Network Based on Residual Train Capacity

    Directory of Open Access Journals (Sweden)

    Fei Dou

    2014-01-01

    Full Text Available Passenger’s optimal path choice is one of the prominent research topics in the field of railway passenger transport organization. More and more different train types are available, increasing path choices from departure to destination for travelers are unstoppable. However, travelers cannot avoid being confused when they hope to choose a perfect travel plan based on various travel time and cost constraints before departure. In this study, railway passenger travel network is constructed based on train timetable. Both the generalized cost function we developed and the residual train capacity are considered to be the foundation of path searching procedure. The railway passenger travel network topology is analyzed based on residual train capacity. Considering the total travel time, the total travel cost, and the total number of passengers, we propose an optimal path searching algorithm based on residual train capacity in railway passenger travel network. Finally, the rationale of the railway passenger travel network and the optimal path generation algorithm are verified positively by case study.

  20. The MARS15-based FermiCORD code system for calculation of the accelerator-induced residual dose

    Science.gov (United States)

    Grebe, A.; Leveling, A.; Lu, T.; Mokhov, N.; Pronskikh, V.

    2018-01-01

    The FermiCORD code system, a set of codes based on MARS15 that calculates the accelerator-induced residual doses at experimental facilities of arbitrary configurations, has been developed. FermiCORD is written in C++ as an add-on to Fortran-based MARS15. The FermiCORD algorithm consists of two stages: 1) simulation of residual doses on contact with the surfaces surrounding the studied location and of radionuclide inventories in the structures surrounding those locations using MARS15, and 2) simulation of the emission of the nuclear decay γ-quanta by the residuals in the activated structures and scoring the prompt doses of these γ-quanta at arbitrary distances from those structures. The FermiCORD code system has been benchmarked against similar algorithms based on other code systems and against experimental data from the CERF facility at CERN, and FermiCORD showed reasonable agreement with these. The code system has been applied for calculation of the residual dose of the target station for the Mu2e experiment and the results have been compared to approximate dosimetric approaches.

  1. Phase-unwrapping algorithm by a rounding-least-squares approach

    Science.gov (United States)

    Juarez-Salazar, Rigoberto; Robledo-Sanchez, Carlos; Guerrero-Sanchez, Fermin

    2014-02-01

    A simple and efficient phase-unwrapping algorithm based on a rounding procedure and a global least-squares minimization is proposed. Instead of processing the gradient of the wrapped phase, this algorithm operates over the gradient of the phase jumps by a robust and noniterative scheme. Thus, the residue-spreading and over-smoothing effects are reduced. The algorithm's performance is compared with four well-known phase-unwrapping methods: minimum cost network flow (MCNF), fast Fourier transform (FFT), quality-guided, and branch-cut. A computer simulation and experimental results show that the proposed algorithm reaches a high-accuracy level than the MCNF method by a low-computing time similar to the FFT phase-unwrapping method. Moreover, since the proposed algorithm is simple, fast, and user-free, it could be used in metrological interferometric and fringe-projection automatic real-time applications.

  2. Approximated affine projection algorithm for feedback cancellation in hearing aids.

    Science.gov (United States)

    Lee, Sangmin; Kim, In-Young; Park, Young-Cheol

    2007-09-01

    We propose an approximated affine projection (AP) algorithm for feedback cancellation in hearing aids. It is based on the conventional approach using the Gauss-Seidel (GS) iteration, but provides more stable convergence behaviour even with small step sizes. In the proposed algorithm, a residue of the weighted error vector, instead of the current error sample, is used to provide stable convergence. A new learning rate control scheme is also applied to the proposed algorithm to prevent signal cancellation and system instability. The new scheme determines step size in proportion to the prediction factor of the input, so that adaptation is inhibited whenever tone-like signals are present in the input. Simulation results verified the efficiency of the proposed algorithm.

  3. Simulation of two-phase flows by domain decomposition

    International Nuclear Information System (INIS)

    Dao, T.H.

    2013-01-01

    This thesis deals with numerical simulations of compressible fluid flows by implicit finite volume methods. Firstly, we studied and implemented an implicit version of the Roe scheme for compressible single-phase and two-phase flows. Thanks to Newton method for solving nonlinear systems, our schemes are conservative. Unfortunately, the resolution of nonlinear systems is very expensive. It is therefore essential to use an efficient algorithm to solve these systems. For large size matrices, we often use iterative methods whose convergence depends on the spectrum. We have studied the spectrum of the linear system and proposed a strategy, called Scaling, to improve the condition number of the matrix. Combined with the classical ILU pre-conditioner, our strategy has reduced significantly the GMRES iterations for local systems and the computation time. We also show some satisfactory results for low Mach-number flows using the implicit centered scheme. We then studied and implemented a domain decomposition method for compressible fluid flows. We have proposed a new interface variable which makes the Schur complement method easy to build and allows us to treat diffusion terms. Using GMRES iterative solver rather than Richardson for the interface system also provides a better performance compared to other methods. We can also decompose the computational domain into any number of sub-domains. Moreover, the Scaling strategy for the interface system has improved the condition number of the matrix and reduced the number of GMRES iterations. In comparison with the classical distributed computing, we have shown that our method is more robust and efficient. (author) [fr

  4. Seismic noise attenuation using an online subspace tracking algorithm

    Science.gov (United States)

    Zhou, Yatong; Li, Shuhua; Zhang, Dong; Chen, Yangkang

    2018-02-01

    We propose a new low-rank based noise attenuation method using an efficient algorithm for tracking subspaces from highly corrupted seismic observations. The subspace tracking algorithm requires only basic linear algebraic manipulations. The algorithm is derived by analysing incremental gradient descent on the Grassmannian manifold of subspaces. When the multidimensional seismic data are mapped to a low-rank space, the subspace tracking algorithm can be directly applied to the input low-rank matrix to estimate the useful signals. Since the subspace tracking algorithm is an online algorithm, it is more robust to random noise than traditional truncated singular value decomposition (TSVD) based subspace tracking algorithm. Compared with the state-of-the-art algorithms, the proposed denoising method can obtain better performance. More specifically, the proposed method outperforms the TSVD-based singular spectrum analysis method in causing less residual noise and also in saving half of the computational cost. Several synthetic and field data examples with different levels of complexities demonstrate the effectiveness and robustness of the presented algorithm in rejecting different types of noise including random noise, spiky noise, blending noise, and coherent noise.

  5. Lifetime and residual strength of materials

    DEFF Research Database (Denmark)

    Nielsen, Lauge Fuglsang

    1997-01-01

    of load amplitude, load average, fractional time under maximum load, and load frequency.The analysis includes prediction of residual strength (re-cycle strength) during the process of load cycling. It is concluded that number of cycles to failure is a very poor design criterion. It is demonstrated how...... the theory developed can be generalized also to consider non-harmonic load variations.Algorithms are presented for design purposes which may be suggested as qualified alternatives to the Palmgren-Miner's methods normally used in fatigue analysis of materials under arbitrary load variations. Prediction...

  6. A fast algorithm for 3D azimuthally anisotropic velocity scan

    KAUST Repository

    Hu, Jingwei; Fomel, Sergey; Ying, Lexing

    2014-01-01

    © 2014 European Association of Geoscientists & Engineers. The conventional velocity scan can be computationally expensive for large-scale seismic data sets, particularly when the presence of anisotropy requires multiparameter scanning. We introduce a fast algorithm for 3D azimuthally anisotropic velocity scan by generalizing the previously proposed 2D butterfly algorithm for hyperbolic Radon transforms. To compute semblance in a two-parameter residual moveout domain, the numerical complexity of our algorithm is roughly O(N3logN) as opposed to O(N5) of the straightforward velocity scan, with N being the representative of the number of points in a particular dimension of either data space or parameter space. Synthetic and field data examples demonstrate the superior efficiency of the proposed algorithm.

  7. A fast algorithm for 3D azimuthally anisotropic velocity scan

    KAUST Repository

    Hu, Jingwei

    2014-11-11

    © 2014 European Association of Geoscientists & Engineers. The conventional velocity scan can be computationally expensive for large-scale seismic data sets, particularly when the presence of anisotropy requires multiparameter scanning. We introduce a fast algorithm for 3D azimuthally anisotropic velocity scan by generalizing the previously proposed 2D butterfly algorithm for hyperbolic Radon transforms. To compute semblance in a two-parameter residual moveout domain, the numerical complexity of our algorithm is roughly O(N3logN) as opposed to O(N5) of the straightforward velocity scan, with N being the representative of the number of points in a particular dimension of either data space or parameter space. Synthetic and field data examples demonstrate the superior efficiency of the proposed algorithm.

  8. Residual nilpotence and residual solubility of groups

    International Nuclear Information System (INIS)

    Mikhailov, R V

    2005-01-01

    The properties of the residual nilpotence and the residual solubility of groups are studied. The main objects under investigation are the class of residually nilpotent groups such that each central extension of these groups is also residually nilpotent and the class of residually soluble groups such that each Abelian extension of these groups is residually soluble. Various examples of groups not belonging to these classes are constructed by homological methods and methods of the theory of modules over group rings. Several applications of the theory under consideration are presented and problems concerning the residual nilpotence of one-relator groups are considered.

  9. An OMIC biomarker detection algorithm TriVote and its application in methylomic biomarker detection.

    Science.gov (United States)

    Xu, Cheng; Liu, Jiamei; Yang, Weifeng; Shu, Yayun; Wei, Zhipeng; Zheng, Weiwei; Feng, Xin; Zhou, Fengfeng

    2018-04-01

    Transcriptomic and methylomic patterns represent two major OMIC data sources impacted by both inheritable genetic information and environmental factors, and have been widely used as disease diagnosis and prognosis biomarkers. Modern transcriptomic and methylomic profiling technologies detect the status of tens of thousands or even millions of probing residues in the human genome, and introduce a major computational challenge for the existing feature selection algorithms. This study proposes a three-step feature selection algorithm, TriVote, to detect a subset of transcriptomic or methylomic residues with highly accurate binary classification performance. TriVote outperforms both filter and wrapper feature selection algorithms with both higher classification accuracy and smaller feature number on 17 transcriptomes and two methylomes. Biological functions of the methylome biomarkers detected by TriVote were discussed for their disease associations. An easy-to-use Python package is also released to facilitate the further applications.

  10. A Multi-Attribute Pheromone Ant Secure Routing Algorithm Based on Reputation Value for Sensor Networks

    Directory of Open Access Journals (Sweden)

    Lin Zhang

    2017-03-01

    Full Text Available With the development of wireless sensor networks, certain network problems have become more prominent, such as limited node resources, low data transmission security, and short network life cycles. To solve these problems effectively, it is important to design an efficient and trusted secure routing algorithm for wireless sensor networks. Traditional ant-colony optimization algorithms exhibit only local convergence, without considering the residual energy of the nodes and many other problems. This paper introduces a multi-attribute pheromone ant secure routing algorithm based on reputation value (MPASR. This algorithm can reduce the energy consumption of a network and improve the reliability of the nodes’ reputations by filtering nodes with higher coincidence rates and improving the method used to update the nodes’ communication behaviors. At the same time, the node reputation value, the residual node energy and the transmission delay are combined to formulate a synthetic pheromone that is used in the formula for calculating the random proportion rule in traditional ant-colony optimization to select the optimal data transmission path. Simulation results show that the improved algorithm can increase both the security of data transmission and the quality of routing service.

  11. A Multi-Attribute Pheromone Ant Secure Routing Algorithm Based on Reputation Value for Sensor Networks

    Science.gov (United States)

    Zhang, Lin; Yin, Na; Fu, Xiong; Lin, Qiaomin; Wang, Ruchuan

    2017-01-01

    With the development of wireless sensor networks, certain network problems have become more prominent, such as limited node resources, low data transmission security, and short network life cycles. To solve these problems effectively, it is important to design an efficient and trusted secure routing algorithm for wireless sensor networks. Traditional ant-colony optimization algorithms exhibit only local convergence, without considering the residual energy of the nodes and many other problems. This paper introduces a multi-attribute pheromone ant secure routing algorithm based on reputation value (MPASR). This algorithm can reduce the energy consumption of a network and improve the reliability of the nodes’ reputations by filtering nodes with higher coincidence rates and improving the method used to update the nodes’ communication behaviors. At the same time, the node reputation value, the residual node energy and the transmission delay are combined to formulate a synthetic pheromone that is used in the formula for calculating the random proportion rule in traditional ant-colony optimization to select the optimal data transmission path. Simulation results show that the improved algorithm can increase both the security of data transmission and the quality of routing service. PMID:28282894

  12. Density of primes in l-th power residues

    Indian Academy of Sciences (India)

    Given a prime number , a finite set of integers S={a1,…,am} and many -th roots of unity ril,i=1,…,m we study the distribution of primes in Q(l) such that the -th residue symbol of ai with respect to is ril, for all . We find out that this is related to the degree of the extension Q(a1l1,…,a1lm)/Q. We give an algorithm ...

  13. Mid-infrared spectroscopy and multivariate analysis for determination of tetracycline residues in cow's milk

    Directory of Open Access Journals (Sweden)

    Lizeth Mariel Casarrubias-Torres

    2018-01-01

    Full Text Available Mid-infrared spectroscopy and chemometric analysis were tested to determine tetracycline's residues in cow's milk. Cow's milk samples (n = 30 were spiked with tetracycline, chlortetracycline, and oxytetracycline in the range of 10-400 µg/l. Chemometric models to quantify each of the tetracycline's residues were developed by applying Partial Components Regression and Partial Least Squares algorithms. The Soft Independent Modeling of Class Analogy model was used to differentiate between pure milk and milk sample with tetracycline residues. The best models for predicting the levels of these antibiotics were obtained using Partial Least Square 1 algorithm (coefficient of determination between 0.997-0.999 and the standard error of calibration from 1.81 to 2.95. The Soft Independent Modeling of Class Analogy model showed well-separated groups allowing classification of milk samples and milk sample with antibiotics. The obtained results demonstrate the great analytical potential of chemometrics coupled with mid-infrared spectroscopy for the prediction of antibiotic in cow's milk at a concentration of microgram per litre (µg/l. This technique can be used to verify the safety of the milk rapidly and reliably.

  14. Newton-Gauss Algorithm of Robust Weighted Total Least Squares Model

    Directory of Open Access Journals (Sweden)

    WANG Bin

    2015-06-01

    Full Text Available Based on the Newton-Gauss iterative algorithm of weighted total least squares (WTLS, a robust WTLS (RWTLS model is presented. The model utilizes the standardized residuals to construct the weight factor function and the square root of the variance component estimator with robustness is obtained by introducing the median method. Therefore, the robustness in both the observation and structure spaces can be simultaneously achieved. To obtain standardized residuals, the linearly approximate cofactor propagation law is employed to derive the expression of the cofactor matrix of WTLS residuals. The iterative calculation steps for RWTLS are also described. The experiment indicates that the model proposed in this paper exhibits satisfactory robustness for gross errors handling problem of WTLS, the obtained parameters have no significant difference with the results of WTLS without gross errors. Therefore, it is superior to the robust weighted total least squares model directly constructed with residuals.

  15. A polynomial time algorithm for solving the maximum flow problem in directed networks

    International Nuclear Information System (INIS)

    Tlas, M.

    2015-01-01

    An efficient polynomial time algorithm for solving maximum flow problems has been proposed in this paper. The algorithm is basically based on the binary representation of capacities; it solves the maximum flow problem as a sequence of O(m) shortest path problems on residual networks with nodes and m arcs. It runs in O(m"2r) time, where is the smallest integer greater than or equal to log B , and B is the largest arc capacity of the network. A numerical example has been illustrated using this proposed algorithm.(author)

  16. NMR structure calculation for all small molecule ligands and non-standard residues from the PDB Chemical Component Dictionary

    Energy Technology Data Exchange (ETDEWEB)

    Yilmaz, Emel Maden; Güntert, Peter, E-mail: guentert@em.uni-frankfurt.de [Goethe University Frankfurt am Main, Center for Biomolecular Magnetic Resonance, Institute of Biophysical Chemistry (Germany)

    2015-09-15

    An algorithm, CYLIB, is presented for converting molecular topology descriptions from the PDB Chemical Component Dictionary into CYANA residue library entries. The CYANA structure calculation algorithm uses torsion angle molecular dynamics for the efficient computation of three-dimensional structures from NMR-derived restraints. For this, the molecules have to be represented in torsion angle space with rotations around covalent single bonds as the only degrees of freedom. The molecule must be given a tree structure of torsion angles connecting rigid units composed of one or several atoms with fixed relative positions. Setting up CYANA residue library entries therefore involves, besides straightforward format conversion, the non-trivial step of defining a suitable tree structure of torsion angles, and to re-order the atoms in a way that is compatible with this tree structure. This can be done manually for small numbers of ligands but the process is time-consuming and error-prone. An automated method is necessary in order to handle the large number of different potential ligand molecules to be studied in drug design projects. Here, we present an algorithm for this purpose, and show that CYANA structure calculations can be performed with almost all small molecule ligands and non-standard amino acid residues in the PDB Chemical Component Dictionary.

  17. Quadratic residues and non-residues selected topics

    CERN Document Server

    Wright, Steve

    2016-01-01

    This book offers an account of the classical theory of quadratic residues and non-residues with the goal of using that theory as a lens through which to view the development of some of the fundamental methods employed in modern elementary, algebraic, and analytic number theory. The first three chapters present some basic facts and the history of quadratic residues and non-residues and discuss various proofs of the Law of Quadratic Reciprosity in depth, with an emphasis on the six proofs that Gauss published. The remaining seven chapters explore some interesting applications of the Law of Quadratic Reciprocity, prove some results concerning the distribution and arithmetic structure of quadratic residues and non-residues, provide a detailed proof of Dirichlet’s Class-Number Formula, and discuss the question of whether quadratic residues are randomly distributed. The text is a valuable resource for graduate and advanced undergraduate students as well as for mathematicians interested in number theory.

  18. Full waveform inversion in the frequency domain using classified time-domain residual wavefields

    Science.gov (United States)

    Son, Woohyun; Koo, Nam-Hyung; Kim, Byoung-Yeop; Lee, Ho-Young; Joo, Yonghwan

    2017-04-01

    We perform the acoustic full waveform inversion in the frequency domain using residual wavefields that have been separated in the time domain. We sort the residual wavefields in the time domain according to the order of absolute amplitudes. Then, the residual wavefields are separated into several groups in the time domain. To analyze the characteristics of the residual wavefields, we compare the residual wavefields of conventional method with those of our residual separation method. From the residual analysis, the amplitude spectrum obtained from the trace before separation appears to have little energy at the lower frequency bands. However, the amplitude spectrum obtained from our strategy is regularized by the separation process, which means that the low-frequency components are emphasized. Therefore, our method helps to emphasize low-frequency components of residual wavefields. Then, we generate the frequency-domain residual wavefields by taking the Fourier transform of the separated time-domain residual wavefields. With these wavefields, we perform the gradient-based full waveform inversion in the frequency domain using back-propagation technique. Through a comparison of gradient directions, we confirm that our separation method can better describe the sub-salt image than the conventional approach. The proposed method is tested on the SEG/EAGE salt-dome model. The inversion results show that our algorithm is better than the conventional gradient based waveform inversion in the frequency domain, especially for deeper parts of the velocity model.

  19. Texture orientation-based algorithm for detecting infrared maritime targets.

    Science.gov (United States)

    Wang, Bin; Dong, Lili; Zhao, Ming; Wu, Houde; Xu, Wenhai

    2015-05-20

    Infrared maritime target detection is a key technology for maritime target searching systems. However, in infrared maritime images (IMIs) taken under complicated sea conditions, background clutters, such as ocean waves, clouds or sea fog, usually have high intensity that can easily overwhelm the brightness of real targets, which is difficult for traditional target detection algorithms to deal with. To mitigate this problem, this paper proposes a novel target detection algorithm based on texture orientation. This algorithm first extracts suspected targets by analyzing the intersubband correlation between horizontal and vertical wavelet subbands of the original IMI on the first scale. Then the self-adaptive wavelet threshold denoising and local singularity analysis of the original IMI is combined to remove false alarms further. Experiments show that compared with traditional algorithms, this algorithm can suppress background clutter much better and realize better single-frame detection for infrared maritime targets. Besides, in order to guarantee accurate target extraction further, the pipeline-filtering algorithm is adopted to eliminate residual false alarms. The high practical value and applicability of this proposed strategy is backed strongly by experimental data acquired under different environmental conditions.

  20. Determination of potassium concentration in salt water for residual beta radioactivity measurements

    International Nuclear Information System (INIS)

    Suarez-Navarro, J.A.; Pujol, Ll.

    2004-01-01

    High interferences may arise in the determination of potassium concentration in salt water. Several analytical methods were studied to determine which method provided the most accurate measurements of potassium concentration. This study is relevant for radiation protection because the exact amount of potassium in water samples must be known for determinations of residual beta activity concentration. The fitting algorithm of the calibration curve and estimation of uncertainty in potassium determinations were also studied. The reproducibility of the proposed analytical method was tested by internal and external validation. Furthermore, the residual beta activity concentration of several Spanish seawater and brackish river water samples was determined using the proposed method

  1. Natural selection and algorithmic design of mRNA.

    Science.gov (United States)

    Cohen, Barry; Skiena, Steven

    2003-01-01

    Messenger RNA (mRNA) sequences serve as templates for proteins according to the triplet code, in which each of the 4(3) = 64 different codons (sequences of three consecutive nucleotide bases) in RNA either terminate transcription or map to one of the 20 different amino acids (or residues) which build up proteins. Because there are more codons than residues, there is inherent redundancy in the coding. Certain residues (e.g., tryptophan) have only a single corresponding codon, while other residues (e.g., arginine) have as many as six corresponding codons. This freedom implies that the number of possible RNA sequences coding for a given protein grows exponentially in the length of the protein. Thus nature has wide latitude to select among mRNA sequences which are informationally equivalent, but structurally and energetically divergent. In this paper, we explore how nature takes advantage of this freedom and how to algorithmically design structures more energetically favorable than have been built through natural selection. In particular: (1) Natural Selection--we perform the first large-scale computational experiment comparing the stability of mRNA sequences from a variety of organisms to random synonymous sequences which respect the codon preferences of the organism. This experiment was conducted on over 27,000 sequences from 34 microbial species with 36 genomic structures. We provide evidence that in all genomic structures highly stable sequences are disproportionately abundant, and in 19 of 36 cases highly unstable sequences are disproportionately abundant. This suggests that the stability of mRNA sequences is subject to natural selection. (2) Artificial Selection--motivated by these biological results, we examine the algorithmic problem of designing the most stable and unstable mRNA sequences which code for a target protein. We give a polynomial-time dynamic programming solution to the most stable sequence problem (MSSP), which is asymptotically no more complex

  2. Some uses of the symmetric Lanczos algorithm - and why it works!

    Energy Technology Data Exchange (ETDEWEB)

    Druskin, V.L. [Schlumberger-Doll Research, Ridgefield, CT (United States); Greenbaum, A. [Courant Institute of Mathematical Sciences, New York, NY (United States); Knizhnerman, L.A. [Central Geophysical Expedition, Moscow (Russian Federation)

    1996-12-31

    The Lanczos algorithm uses a three-term recurrence to construct an orthonormal basis for the Krylov space corresponding to a symmetric matrix A and a starting vector q{sub 1}. The vectors and recurrence coefficients produced by this algorithm can be used for a number of purposes, including solving linear systems Au = {var_phi} and computing the matrix exponential e{sup -tA}{var_phi}. Although the vectors produced in finite precision arithmetic are not orthogonal, we show why they can still be used effectively for these purposes. The reason is that the 2-norm of the residual is essentially determined by the tridiagonal matrix and the next recurrence coefficient produced by the finite precision Lanczos computation. It follows that if the same tridiagonal matrix and recurrence coefficient are produced by the exact Lanczos algorithm applied to some other problem, then exact arithmetic bounds on the residual for that problem will hold for the finite precision computation. In order to establish exact arithmetic bounds for the different problem, it is necessary to have some information about the eigenvalues of the new coefficient matrix. Here we make use of information already established in the literature, and we also prove a new result for indefinite matrices.

  3. On an algorithm of data compressian under filmless pickup of data from streamer chambers

    International Nuclear Information System (INIS)

    Ososkov, G.A.; Perelygin, S.P.; Prikhod'ko, V.I.; Ton, T.; Chelnokova, V.V.

    1978-01-01

    A primary data compression algorithm is discussed with features feasible loss of accuracy during reconstruction of event geometry. The most effective methods of data compression problem solution are: residual classes calculus and contour following. An approach toWards residual classes calculus operation is suggested, Which consists of two stages of digitized data processing. First, transformation of Cartesian coordination system With 12 13 x2 9 samples along X and Y axes, correspondingly, into a system with 2 8 x2 4 samples. The second stage can be fulfilled in either way: simple sorting of all transformed coordinates of tracks and interference - this algorithm can be implemented in two 2 8 x2 5 bit matrices (per each TV camera) resulting in 13-15-fold compression or track following with calculation of X increment perline at the end of each slice. The second algorithm needs a set of followers which can be implemented in 16x100 bit matrix. It also requires a controller which has some 100 medium IC's. This will enable to achieve 20-30-fold compression of data

  4. The Relationship Between Low-Frequency Motions and Community Structure of Residue Network in Protein Molecules.

    Science.gov (United States)

    Sun, Weitao

    2018-01-01

    The global shape of a protein molecule is believed to be dominant in determining low-frequency deformational motions. However, how structure dynamics relies on residue interactions remains largely unknown. The global residue community structure and the local residue interactions are two important coexisting factors imposing significant effects on low-frequency normal modes. In this work, an algorithm for community structure partition is proposed by integrating Miyazawa-Jernigan empirical potential energy as edge weight. A sensitivity parameter is defined to measure the effect of local residue interaction on low-frequency movement. We show that community structure is a more fundamental feature of residue contact networks. Moreover, we surprisingly find that low-frequency normal mode eigenvectors are sensitive to some local critical residue interaction pairs (CRIPs). A fair amount of CRIPs act as bridges and hold distributed structure components into a unified tertiary structure by bonding nearby communities. Community structure analysis and CRIP detection of 116 catalytic proteins reveal that breaking up of a CRIP can cause low-frequency allosteric movement of a residue at the far side of protein structure. The results imply that community structure and CRIP may be the structural basis for low-frequency motions.

  5. Model-based fault detection algorithm for photovoltaic system monitoring

    KAUST Repository

    Harrou, Fouzi

    2018-02-12

    Reliable detection of faults in PV systems plays an important role in improving their reliability, productivity, and safety. This paper addresses the detection of faults in the direct current (DC) side of photovoltaic (PV) systems using a statistical approach. Specifically, a simulation model that mimics the theoretical performances of the inspected PV system is designed. Residuals, which are the difference between the measured and estimated output data, are used as a fault indicator. Indeed, residuals are used as the input for the Multivariate CUmulative SUM (MCUSUM) algorithm to detect potential faults. We evaluated the proposed method by using data from an actual 20 MWp grid-connected PV system located in the province of Adrar, Algeria.

  6. Optimised ExpTime Tableaux for ℋℐ over Finite Residuated Lattices

    Directory of Open Access Journals (Sweden)

    Jian Huang

    2014-01-01

    Full Text Available This study proposes to adopt a novel tableau reasoning algorithm for the description logic ℋℐ with semantics based on a finite residuated De Morgan lattice. The syntax, semantics, and logical properties of this logic are given, and a sound, complete, and terminating tableaux algorithm for deciding fuzzy ABox consistency and concept satisfiability problem with respect to TBox is presented. Moreover, based on extended and/or completion-forest with a series of sound optimization technique for checking satisfiability with respect to a TBox in the logic, a new optimized ExpTime (complexity-optimal tableau decision procedure is presented here. The experimental evaluation indicates that the optimization techniques we considered result in improved efficiency significantly.

  7. Application of preconditioned conjugate gradient-like methods to reactor kinetics

    International Nuclear Information System (INIS)

    Yang, D.Y.; Chen, G.S.; Chou, H.P.

    1993-01-01

    Several conjugate gradient-like (CG-like) methods are applied to solve the nonsymmetric linear systems of equations derived from the time-dependent two-dimensional two-energy-group neutron diffusion equations by a finite difference approximation. The methods are: the generalized conjugate residual method; the generalized conjugate gradient least-square method; the generalized minimal residual method (GMRES); the conjugate gradient square method; and a variant of bi-conjugate gradient method (Bi-CGSTAB). In order to accelerate these methods, six preconditioning techniques are investigated. Two are based on pointwise incomplete factorization: the incomplete LU (ILU) and the modified incomplete LU (MILU) decompositions; two, based on the block tridiagonal structure of the coefficient matrix, are blockwise and modified blockwise incomplete factorizations, BILU and MBILU; two are the alternating-direction implicit and symmetric successive overrelaxation (SSOR) preconditioners, derived from the basic iterative schemes. Comparisons are made by using CG-like methods combined with different preconditioners to solve a sequence of time-step reactor transient problems. Numerical tests indicate that preconditioned BI-CGSTAB with the preconditioner MBILU requires less CPU time and fewer iterations than other methods. The preconditioned CG-like methods are less sensitive to the time-step size used than the Chebyshev semi-iteration method and line SOR method. The indication is that the CGS, Bi-CGSTAB and GMRES methods are, on average, better than the other methods in reactor kinetics computation and that a good preconditioner is more important than the choice of CG-like methods. The MILU decomposition based on the conventional row-sum criterion has difficulty yielding a convergent solution and an improved version is introduced. (author)

  8. An adjoint method of sensitivity analysis for residual vibrations of structures subject to impacts

    Science.gov (United States)

    Yan, Kun; Cheng, Gengdong

    2018-03-01

    For structures subject to impact loads, the residual vibration reduction is more and more important as the machines become faster and lighter. An efficient sensitivity analysis of residual vibration with respect to structural or operational parameters is indispensable for using a gradient based optimization algorithm, which reduces the residual vibration in either active or passive way. In this paper, an integrated quadratic performance index is used as the measure of the residual vibration, since it globally measures the residual vibration response and its calculation can be simplified greatly with Lyapunov equation. Several sensitivity analysis approaches for performance index were developed based on the assumption that the initial excitations of residual vibration were given and independent of structural design. Since the resulting excitations by the impact load often depend on structural design, this paper aims to propose a new efficient sensitivity analysis method for residual vibration of structures subject to impacts to consider the dependence. The new method is developed by combining two existing methods and using adjoint variable approach. Three numerical examples are carried out and demonstrate the accuracy of the proposed method. The numerical results show that the dependence of initial excitations on structural design variables may strongly affects the accuracy of sensitivities.

  9. Fault Severity Estimation of Rotating Machinery Based on Residual Signals

    Directory of Open Access Journals (Sweden)

    Fan Jiang

    2012-01-01

    Full Text Available Fault severity estimation is an important part of a condition-based maintenance system, which can monitor the performance of an operation machine and enhance its level of safety. In this paper, a novel method based on statistical property and residual signals is developed for estimating the fault severity of rotating machinery. The fast Fourier transformation (FFT is applied to extract the so-called multifrequency-band energy (MFBE from the vibration signals of rotating machinery with different fault severity levels in the first stage. Usually these features of the working conditions with different fault sensitivities are different. Therefore a sensitive features-selecting algorithm is defined to construct the feature matrix and calculate the statistic parameter (mean in the second stage. In the last stage, the residual signals computed by the zero space vector are used to estimate the fault severity. Simulation and experimental results reveal that the proposed method based on statistics and residual signals is effective and feasible for estimating the severity of a rotating machine fault.

  10. Reliable and Efficient Parallel Processing Algorithms and Architectures for Modern Signal Processing. Ph.D. Thesis

    Science.gov (United States)

    Liu, Kuojuey Ray

    1990-01-01

    Least-squares (LS) estimations and spectral decomposition algorithms constitute the heart of modern signal processing and communication problems. Implementations of recursive LS and spectral decomposition algorithms onto parallel processing architectures such as systolic arrays with efficient fault-tolerant schemes are the major concerns of this dissertation. There are four major results in this dissertation. First, we propose the systolic block Householder transformation with application to the recursive least-squares minimization. It is successfully implemented on a systolic array with a two-level pipelined implementation at the vector level as well as at the word level. Second, a real-time algorithm-based concurrent error detection scheme based on the residual method is proposed for the QRD RLS systolic array. The fault diagnosis, order degraded reconfiguration, and performance analysis are also considered. Third, the dynamic range, stability, error detection capability under finite-precision implementation, order degraded performance, and residual estimation under faulty situations for the QRD RLS systolic array are studied in details. Finally, we propose the use of multi-phase systolic algorithms for spectral decomposition based on the QR algorithm. Two systolic architectures, one based on triangular array and another based on rectangular array, are presented for the multiphase operations with fault-tolerant considerations. Eigenvectors and singular vectors can be easily obtained by using the multi-pase operations. Performance issues are also considered.

  11. Accurate prediction of hot spot residues through physicochemical characteristics of amino acid sequences

    KAUST Repository

    Chen, Peng; Li, Jinyan; Limsoon, Wong; Kuwahara, Hiroyuki; Huang, Jianhua Z.; Gao, Xin

    2013-01-01

    Hot spot residues of proteins are fundamental interface residues that help proteins perform their functions. Detecting hot spots by experimental methods is costly and time-consuming. Sequential and structural information has been widely used in the computational prediction of hot spots. However, structural information is not always available. In this article, we investigated the problem of identifying hot spots using only physicochemical characteristics extracted from amino acid sequences. We first extracted 132 relatively independent physicochemical features from a set of the 544 properties in AAindex1, an amino acid index database. Each feature was utilized to train a classification model with a novel encoding schema for hot spot prediction by the IBk algorithm, an extension of the K-nearest neighbor algorithm. The combinations of the individual classifiers were explored and the classifiers that appeared frequently in the top performing combinations were selected. The hot spot predictor was built based on an ensemble of these classifiers and to work in a voting manner. Experimental results demonstrated that our method effectively exploited the feature space and allowed flexible weights of features for different queries. On the commonly used hot spot benchmark sets, our method significantly outperformed other machine learning algorithms and state-of-the-art hot spot predictors. The program is available at http://sfb.kaust.edu.sa/pages/software.aspx. © 2013 Wiley Periodicals, Inc.

  12. Accurate prediction of hot spot residues through physicochemical characteristics of amino acid sequences

    KAUST Repository

    Chen, Peng

    2013-07-23

    Hot spot residues of proteins are fundamental interface residues that help proteins perform their functions. Detecting hot spots by experimental methods is costly and time-consuming. Sequential and structural information has been widely used in the computational prediction of hot spots. However, structural information is not always available. In this article, we investigated the problem of identifying hot spots using only physicochemical characteristics extracted from amino acid sequences. We first extracted 132 relatively independent physicochemical features from a set of the 544 properties in AAindex1, an amino acid index database. Each feature was utilized to train a classification model with a novel encoding schema for hot spot prediction by the IBk algorithm, an extension of the K-nearest neighbor algorithm. The combinations of the individual classifiers were explored and the classifiers that appeared frequently in the top performing combinations were selected. The hot spot predictor was built based on an ensemble of these classifiers and to work in a voting manner. Experimental results demonstrated that our method effectively exploited the feature space and allowed flexible weights of features for different queries. On the commonly used hot spot benchmark sets, our method significantly outperformed other machine learning algorithms and state-of-the-art hot spot predictors. The program is available at http://sfb.kaust.edu.sa/pages/software.aspx. © 2013 Wiley Periodicals, Inc.

  13. Accurate prediction of hot spot residues through physicochemical characteristics of amino acid sequences.

    Science.gov (United States)

    Chen, Peng; Li, Jinyan; Wong, Limsoon; Kuwahara, Hiroyuki; Huang, Jianhua Z; Gao, Xin

    2013-08-01

    Hot spot residues of proteins are fundamental interface residues that help proteins perform their functions. Detecting hot spots by experimental methods is costly and time-consuming. Sequential and structural information has been widely used in the computational prediction of hot spots. However, structural information is not always available. In this article, we investigated the problem of identifying hot spots using only physicochemical characteristics extracted from amino acid sequences. We first extracted 132 relatively independent physicochemical features from a set of the 544 properties in AAindex1, an amino acid index database. Each feature was utilized to train a classification model with a novel encoding schema for hot spot prediction by the IBk algorithm, an extension of the K-nearest neighbor algorithm. The combinations of the individual classifiers were explored and the classifiers that appeared frequently in the top performing combinations were selected. The hot spot predictor was built based on an ensemble of these classifiers and to work in a voting manner. Experimental results demonstrated that our method effectively exploited the feature space and allowed flexible weights of features for different queries. On the commonly used hot spot benchmark sets, our method significantly outperformed other machine learning algorithms and state-of-the-art hot spot predictors. The program is available at http://sfb.kaust.edu.sa/pages/software.aspx. Copyright © 2013 Wiley Periodicals, Inc.

  14. Efficient identification of critical residues based only on protein structure by network analysis.

    Directory of Open Access Journals (Sweden)

    Michael P Cusack

    2007-05-01

    Full Text Available Despite the increasing number of published protein structures, and the fact that each protein's function relies on its three-dimensional structure, there is limited access to automatic programs used for the identification of critical residues from the protein structure, compared with those based on protein sequence. Here we present a new algorithm based on network analysis applied exclusively on protein structures to identify critical residues. Our results show that this method identifies critical residues for protein function with high reliability and improves automatic sequence-based approaches and previous network-based approaches. The reliability of the method depends on the conformational diversity screened for the protein of interest. We have designed a web site to give access to this software at http://bis.ifc.unam.mx/jamming/. In summary, a new method is presented that relates critical residues for protein function with the most traversed residues in networks derived from protein structures. A unique feature of the method is the inclusion of the conformational diversity of proteins in the prediction, thus reproducing a basic feature of the structure/function relationship of proteins.

  15. A multilevel method for conductive-radiative heat transfer

    Energy Technology Data Exchange (ETDEWEB)

    Banoczi, J.M.; Kelley, C.T. [North Carolina State Univ., Raleigh, NC (United States)

    1996-12-31

    We present a fast multilevel algorithm for the solution of a system of nonlinear integro-differential equations that model steady-state combined radiative-conductive heat transfer. The equations can be formulated as a compact fixed point problem with a fixed point map that requires both a solution of the linear transport equation and the linear heat equation for its evaluation. We use fast transport solvers developed by the second author, to construct an efficient evaluation of the fixed point map and then apply the Atkinson-Brakhage, method, with Newton-GMRES as the coarse mesh solver, to the full nonlinear system.

  16. On performance of Krylov smoothing for fully-coupled AMG preconditioners for VMS resistive MHD

    Energy Technology Data Exchange (ETDEWEB)

    Lin, Paul T. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Shadid, John N. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Univ. of New Mexico, Albuquerque, NM (United States). Department of Mathematics and Statistics,; Tsuji, Paul H. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-11-01

    Here, this study explores the performance and scaling of a GMRES Krylov method employed as a smoother for an algebraic multigrid (AMG) preconditioned Newton- Krylov solution approach applied to a fully-implicit variational multiscale (VMS) nite element (FE) resistive magnetohydrodynamics (MHD) formulation. In this context a Newton iteration is used for the nonlinear system and a Krylov (GMRES) method is employed for the linear subsystems. The efficiency of this approach is critically dependent on the scalability and performance of the AMG preconditioner for the linear solutions and the performance of the smoothers play a critical role. Krylov smoothers are considered in an attempt to reduce the time and memory requirements of existing robust smoothers based on additive Schwarz domain decomposition (DD) with incomplete LU factorization solves on each subdomain. Three time dependent resistive MHD test cases are considered to evaluate the method. The results demonstrate that the GMRES smoother can be faster due to a decrease in the preconditioner setup time and a reduction in outer GMRESR solver iterations, and requires less memory (typically 35% less memory for global GMRES smoother) than the DD ILU smoother.

  17. An improved algorithm for MFR fragment assembly

    International Nuclear Information System (INIS)

    Kontaxis, Georg

    2012-01-01

    A method for generating protein backbone models from backbone only NMR data is presented, which is based on molecular fragment replacement (MFR). In a first step, the PDB database is mined for homologous peptide fragments using experimental backbone-only data i.e. backbone chemical shifts (CS) and residual dipolar couplings (RDC). Second, this fragment library is refined against the experimental restraints. Finally, the fragments are assembled into a protein backbone fold using a rigid body docking algorithm using the RDCs as restraints. For improved performance, backbone nuclear Overhauser effects (NOEs) may be included at that stage. Compared to previous implementations of MFR-derived structure determination protocols this model-building algorithm offers improved stability and reliability. Furthermore, relative to CS-ROSETTA based methods, it provides faster performance and straightforward implementation with the option to easily include further types of restraints and additional energy terms.

  18. An efficient randomized algorithm for contact-based NMR backbone resonance assignment.

    Science.gov (United States)

    Kamisetty, Hetunandan; Bailey-Kellogg, Chris; Pandurangan, Gopal

    2006-01-15

    Backbone resonance assignment is a critical bottleneck in studies of protein structure, dynamics and interactions by nuclear magnetic resonance (NMR) spectroscopy. A minimalist approach to assignment, which we call 'contact-based', seeks to dramatically reduce experimental time and expense by replacing the standard suite of through-bond experiments with the through-space (nuclear Overhauser enhancement spectroscopy, NOESY) experiment. In the contact-based approach, spectral data are represented in a graph with vertices for putative residues (of unknown relation to the primary sequence) and edges for hypothesized NOESY interactions, such that observed spectral peaks could be explained if the residues were 'close enough'. Due to experimental ambiguity, several incorrect edges can be hypothesized for each spectral peak. An assignment is derived by identifying consistent patterns of edges (e.g. for alpha-helices and beta-sheets) within a graph and by mapping the vertices to the primary sequence. The key algorithmic challenge is to be able to uncover these patterns even when they are obscured by significant noise. This paper develops, analyzes and applies a novel algorithm for the identification of polytopes representing consistent patterns of edges in a corrupted NOESY graph. Our randomized algorithm aggregates simplices into polytopes and fixes inconsistencies with simple local modifications, called rotations, that maintain most of the structure already uncovered. In characterizing the effects of experimental noise, we employ an NMR-specific random graph model in proving that our algorithm gives optimal performance in expected polynomial time, even when the input graph is significantly corrupted. We confirm this analysis in simulation studies with graphs corrupted by up to 500% noise. Finally, we demonstrate the practical application of the algorithm on several experimental beta-sheet datasets. Our approach is able to eliminate a large majority of noise edges and to

  19. On structure-exploiting trust-region regularized nonlinear least squares algorithms for neural-network learning.

    Science.gov (United States)

    Mizutani, Eiji; Demmel, James W

    2003-01-01

    This paper briefly introduces our numerical linear algebra approaches for solving structured nonlinear least squares problems arising from 'multiple-output' neural-network (NN) models. Our algorithms feature trust-region regularization, and exploit sparsity of either the 'block-angular' residual Jacobian matrix or the 'block-arrow' Gauss-Newton Hessian (or Fisher information matrix in statistical sense) depending on problem scale so as to render a large class of NN-learning algorithms 'efficient' in both memory and operation costs. Using a relatively large real-world nonlinear regression application, we shall explain algorithmic strengths and weaknesses, analyzing simulation results obtained by both direct and iterative trust-region algorithms with two distinct NN models: 'multilayer perceptrons' (MLP) and 'complementary mixtures of MLP-experts' (or neuro-fuzzy modular networks).

  20. Residual motion compensation in ECG-gated interventional cardiac vasculature reconstruction

    Science.gov (United States)

    Schwemmer, C.; Rohkohl, C.; Lauritsch, G.; Müller, K.; Hornegger, J.

    2013-06-01

    Three-dimensional reconstruction of cardiac vasculature from angiographic C-arm CT (rotational angiography) data is a major challenge. Motion artefacts corrupt image quality, reducing usability for diagnosis and guidance. Many state-of-the-art approaches depend on retrospective ECG-gating of projection data for image reconstruction. A trade-off has to be made regarding the size of the ECG-gating window. A large temporal window is desirable to avoid undersampling. However, residual motion will occur in a large window, causing motion artefacts. We present an algorithm to correct for residual motion. Our approach is based on a deformable 2D-2D registration between the forward projection of an initial, ECG-gated reconstruction, and the original projection data. The approach is fully automatic and does not require any complex segmentation of vasculature, or landmarks. The estimated motion is compensated for during the backprojection step of a subsequent reconstruction. We evaluated the method using the publicly available CAVAREV platform and on six human clinical datasets. We found a better visibility of structure, reduced motion artefacts, and increased sharpness of the vessels in the compensated reconstructions compared to the initial reconstructions. At the time of writing, our algorithm outperforms the leading result of the CAVAREV ranking list. For the clinical datasets, we found an average reduction of motion artefacts by 13 ± 6%. Vessel sharpness was improved by 25 ± 12% on average.

  1. Use of a dynamic grid adaptation in the asymmetric weighted residual method

    International Nuclear Information System (INIS)

    Graf, V.; Romstedt, P.; Werner, W.

    1986-01-01

    A dynamic grid adaptive method has been developed for use with the asymmetric weighted residual method. The method automatically adapts the number and position of the spatial mesh points as the solution of hyperbolic or parabolic vector partial differential equations progresses in time. The mesh selection algorithm is based on the minimization of the L 2 norm of the spatial discretization error. The method permits the accurate calculation of the evolution of inhomogeneities, like wave fronts, shock layers, and other sharp transitions, while generally using a coarse computational grid. The number of required mesh points is significantly reduced, relative to a fixed Eulerian grid. Since the mesh selection algorithm is computationally inexpensive, a corresponding reduction of computing time results

  2. Comparison of Nonequilibrium Solution Algorithms Applied to Chemically Stiff Hypersonic Flows

    Science.gov (United States)

    Palmer, Grant; Venkatapathy, Ethiraj

    1995-01-01

    Three solution algorithms, explicit under-relaxation, point implicit, and lower-upper symmetric Gauss-Seidel, are used to compute nonequilibrium flow around the Apollo 4 return capsule at the 62-km altitude point in its descent trajectory. By varying the Mach number, the efficiency and robustness of the solution algorithms were tested for different levels of chemical stiffness.The performance of the solution algorithms degraded as the Mach number and stiffness of the flow increased. At Mach 15 and 30, the lower-upper symmetric Gauss-Seidel method produces an eight order of magnitude drop in the energy residual in one-third to one-half the Cray C-90 computer time as compared to the point implicit and explicit under-relaxation methods. The explicit under-relaxation algorithm experienced convergence difficulties at Mach 30 and above. At Mach 40 the performance of the lower-upper symmetric Gauss-Seidel algorithm deteriorates to the point that it is out performed by the point implicit method. The effects of the viscous terms are investigated. Grid dependency questions are explored.

  3. Residual stresses

    International Nuclear Information System (INIS)

    Sahotra, I.M.

    2006-01-01

    The principal effect of unloading a material strained into the plastic range is to create a permanent set (plastic deformation), which if restricted somehow, gives rise to a system of self-balancing within the same member or reaction balanced by other members of the structure., known as residual stresses. These stresses stay there as locked-in stresses, in the body or a part of it in the absence of any external loading. Residual stresses are induced during hot-rolling and welding differential cooling, cold-forming and extruding: cold straightening and spot heating, fabrication and forced fitting of components constraining the structure to a particular geometry. The areas which cool more quickly develop residual compressive stresses, while the slower cooling areas develop residual tensile stresses, and a self-balancing or reaction balanced system of residual stresses is formed. The phenomenon of residual stresses is the most challenging in its application in surface modification techniques determining endurance mechanism against fracture and fatigue failures. This paper discusses the mechanism of residual stresses, that how the residual stresses are fanned and what their behavior is under the action of external forces. Such as in the case of a circular bar under limit torque, rectangular beam under limt moment, reclaiming of shafts welds and peening etc. (author)

  4. Research on optimization algorithms and applications. Progress report, July 1, 1997 - June 30, 1998

    International Nuclear Information System (INIS)

    Zhang, Y.

    1998-06-01

    The project had a productive first year with five technical reports produced or in preparation. The major works accomplished include the following: (1) the authors have conducted investigations on algorithms for global optimization of zero-to-small residual least squares problems, which is important in x-ray crystallography. Numerical results have demonstrated the effectiveness of their algorithms on many test problems; (2) they have investigated a fundamental question concerning the behavior of Newton's method on equivalent systems. This investigation provides guidance to the proper choice of formulations in solving some optimization problems; (3) they have developed and tested new and promising interior-point algorithms for the maximum-volume ellipsoid problem

  5. The structure of the polynomials in preconditioned BiCG algorithms and the switching direction of preconditioned systems

    OpenAIRE

    Itoh, Shoji; Sugihara, Masaaki

    2016-01-01

    We present a theorem that defines the direction of a preconditioned system for the bi-conjugate gradient (BiCG) method, and we extend it to preconditioned bi-Lanczos-type algorithms. We show that the direction of a preconditioned system is switched by construction and by the settings of the initial shadow residual vector. We analyze and compare the polynomial structures of four preconditioned BiCG algorithms.

  6. Comparison of multihardware parallel implementations for a phase unwrapping algorithm

    Science.gov (United States)

    Hernandez-Lopez, Francisco Javier; Rivera, Mariano; Salazar-Garibay, Adan; Legarda-Sáenz, Ricardo

    2018-04-01

    Phase unwrapping is an important problem in the areas of optical metrology, synthetic aperture radar (SAR) image analysis, and magnetic resonance imaging (MRI) analysis. These images are becoming larger in size and, particularly, the availability and need for processing of SAR and MRI data have increased significantly with the acquisition of remote sensing data and the popularization of magnetic resonators in clinical diagnosis. Therefore, it is important to develop faster and accurate phase unwrapping algorithms. We propose a parallel multigrid algorithm of a phase unwrapping method named accumulation of residual maps, which builds on a serial algorithm that consists of the minimization of a cost function; minimization achieved by means of a serial Gauss-Seidel kind algorithm. Our algorithm also optimizes the original cost function, but unlike the original work, our algorithm is a parallel Jacobi class with alternated minimizations. This strategy is known as the chessboard type, where red pixels can be updated in parallel at same iteration since they are independent. Similarly, black pixels can be updated in parallel in an alternating iteration. We present parallel implementations of our algorithm for different parallel multicore architecture such as CPU-multicore, Xeon Phi coprocessor, and Nvidia graphics processing unit. In all the cases, we obtain a superior performance of our parallel algorithm when compared with the original serial version. In addition, we present a detailed comparative performance of the developed parallel versions.

  7. Residual stresses

    International Nuclear Information System (INIS)

    Macherauch, E.

    1978-01-01

    Residual stresses are stresses which exist in a material without the influence of external powers and moments. They come into existence when the volume of a material constantly changes its form as a consequence of mechanical, thermal, and/or chemical processes and is hindered by neighbouring volumes. Bodies with residual stress are in mechanical balance. These residual stresses can be manifested by means of all mechanical interventions disturbing this balance. Acoustical, optical, radiological, and magnetical methods involving material changes caused by residual stress can also serve for determining residual stress. Residual stresses have an ambivalent character. In technical practice, they are feared and liked at the same time. They cause trouble because they can be the cause for unexpected behaviour of construction elements. They are feared since they can cause failure, in the worst case with catastrophical consequences. They are appreciated, on the other hand, because, in many cases, they can contribute to improvements of the material behaviour under certain circumstances. But they are especially liked for their giving convenient and (this is most important) mostly uncontrollable explanations. For only in very few cases we have enough knowledge and possibilities for the objective evaluation of residual stresses. (orig.) [de

  8. Optimizing Biomass Feedstock Logistics for Forest Residue Processing and Transportation on a Tree-Shaped Road Network

    Directory of Open Access Journals (Sweden)

    Hee Han

    2018-03-01

    Full Text Available An important task in forest residue recovery operations is to select the most cost-efficient feedstock logistics system for a given distribution of residue piles, road access, and available machinery. Notable considerations include inaccessibility of treatment units to large chip vans and frequent, long-distance mobilization of forestry equipment required to process dispersed residues. In this study, we present optimized biomass feedstock logistics on a tree-shaped road network that take into account the following options: (1 grinding residues at the site of treatment and forwarding ground residues either directly to bioenergy facility or to a concentration yard where they are transshipped to large chip vans, (2 forwarding residues to a concentration yard where they are stored and ground directly into chip vans, and (3 forwarding residues to a nearby grinder location and forwarding the ground materials. A mixed-integer programming model coupled with a network algorithm was developed to solve the problem. The model was applied to recovery operations on a study site in Colorado, USA, and the optimal solution reduced the cost of logistics up to 11% compared to the conventional system. This is an important result because this cost reduction propagates downstream through the biomass supply chain, reducing production costs for bioenergy and bioproducts.

  9. Evaluation of residue-residue contact predictions in CASP9

    KAUST Repository

    Monastyrskyy, Bohdan

    2011-01-01

    This work presents the results of the assessment of the intramolecular residue-residue contact predictions submitted to CASP9. The methodology for the assessment does not differ from that used in previous CASPs, with two basic evaluation measures being the precision in recognizing contacts and the difference between the distribution of distances in the subset of predicted contact pairs versus all pairs of residues in the structure. The emphasis is placed on the prediction of long-range contacts (i.e., contacts between residues separated by at least 24 residues along sequence) in target proteins that cannot be easily modeled by homology. Although there is considerable activity in the field, the current analysis reports no discernable progress since CASP8.

  10. Toward an optimal solver for time-spectral fluid-dynamic and aeroelastic solutions on unstructured meshes

    Science.gov (United States)

    Mundis, Nathan L.; Mavriplis, Dimitri J.

    2017-09-01

    The time-spectral method applied to the Euler and coupled aeroelastic equations theoretically offers significant computational savings for purely periodic problems when compared to standard time-implicit methods. However, attaining superior efficiency with time-spectral methods over traditional time-implicit methods hinges on the ability rapidly to solve the large non-linear system resulting from time-spectral discretizations which become larger and stiffer as more time instances are employed or the period of the flow becomes especially short (i.e. the maximum resolvable wave-number increases). In order to increase the efficiency of these solvers, and to improve robustness, particularly for large numbers of time instances, the Generalized Minimal Residual Method (GMRES) is used to solve the implicit linear system over all coupled time instances. The use of GMRES as the linear solver makes time-spectral methods more robust, allows them to be applied to a far greater subset of time-accurate problems, including those with a broad range of harmonic content, and vastly improves the efficiency of time-spectral methods. In previous work, a wave-number independent preconditioner that mitigates the increased stiffness of the time-spectral method when applied to problems with large resolvable wave numbers has been developed. This preconditioner, however, directly inverts a large matrix whose size increases in proportion to the number of time instances. As a result, the computational time of this method scales as the cube of the number of time instances. In the present work, this preconditioner has been reworked to take advantage of an approximate-factorization approach that effectively decouples the spatial and temporal systems. Once decoupled, the time-spectral matrix can be inverted in frequency space, where it has entries only on the main diagonal and therefore can be inverted quite efficiently. This new GMRES/preconditioner combination is shown to be over an order of

  11. Detecting an atomic clock frequency anomaly using an adaptive Kalman filter algorithm

    Science.gov (United States)

    Song, Huijie; Dong, Shaowu; Wu, Wenjun; Jiang, Meng; Wang, Weixiong

    2018-06-01

    The abnormal frequencies of an atomic clock mainly include frequency jump and frequency drift jump. Atomic clock frequency anomaly detection is a key technique in time-keeping. The Kalman filter algorithm, as a linear optimal algorithm, has been widely used in real-time detection for abnormal frequency. In order to obtain an optimal state estimation, the observation model and dynamic model of the Kalman filter algorithm should satisfy Gaussian white noise conditions. The detection performance is degraded if anomalies affect the observation model or dynamic model. The idea of the adaptive Kalman filter algorithm, applied to clock frequency anomaly detection, uses the residuals given by the prediction for building ‘an adaptive factor’ the prediction state covariance matrix is real-time corrected by the adaptive factor. The results show that the model error is reduced and the detection performance is improved. The effectiveness of the algorithm is verified by the frequency jump simulation, the frequency drift jump simulation and the measured data of the atomic clock by using the chi-square test.

  12. Multichannel Filtered-X Error Coded Affine Projection-Like Algorithm with Evolving Order

    Directory of Open Access Journals (Sweden)

    J. G. Avalos

    2017-01-01

    Full Text Available Affine projection (AP algorithms are commonly used to implement active noise control (ANC systems because they provide fast convergence. However, their high computational complexity can restrict their use in certain practical applications. The Error Coded Affine Projection-Like (ECAP-L algorithm has been proposed to reduce the computational burden while maintaining the speed of AP, but no version of this algorithm has been derived for active noise control, for which the adaptive structures are very different from those of other configurations. In this paper, we introduce a version of the ECAP-L for single-channel and multichannel ANC systems. The proposed algorithm is implemented using the conventional filtered-x scheme, which incurs a lower computational cost than the modified filtered-x structure, especially for multichannel systems. Furthermore, we present an evolutionary method that dynamically decreases the projection order in order to reduce the dimensions of the matrix used in the algorithm’s computations. Experimental results demonstrate that the proposed algorithm yields a convergence speed and a final residual error similar to those of AP algorithms. Moreover, it achieves meaningful computational savings, leading to simpler hardware implementation of real-time ANC applications.

  13. ABOUT COMPLEX OPERATIONS IN NON-POSITIONAL RESIDUE NUMBER SYSTEM

    Directory of Open Access Journals (Sweden)

    Yu. D. Polissky

    2016-04-01

    Full Text Available Purpose. The purpose of this work is the theoretical substantiation of methods for increased efficiency of execution of difficult, so-called not modular, operations in non-positional residue number system for which it is necessary to know operand digits for all grade levels. Methodology. To achieve the target the numbers are presented in odd module system, while the result of the operation is determined on the basis of establishing the operand parity. The parity is determined by finding the sum modulo for the values of the number positional characteristics for all of its modules. Algorithm of position characteristics includes two types of iteration. The first iteration is to move from this number to a smaller number, in which the remains of one or more modules are equal to zero. This is achieved by subtracting out of all the residues the value of one of them. The second iteration is to move from this number to a smaller number due to exclusion of modules, which residues are zero, by dividing this number by the product of these modules. Iterations are performed until the residues of one, some or all of the modules equal to zero and other modules are excluded. The proposed method is distinguished by its simplicity and allows you to obtain the result of the operation quickly. Findings. There are obtained rather simple solutions of not modular operations for definition of outputs beyond the range of the result of adding or subtracting pairs of numbers, comparing pairs of numbers, determining the number belonging to the specific half of the range, defining parity of numbers presented in non-positional residue number system. Originality. The work offered the new effective approaches to solve the non-modular operations of the non-positional residue number system. It seems appropriate to consider these approaches as research areas to enhance the effectiveness of the modular calculation. Practical value. The above solutions have high performance and can

  14. Residual motion compensation in ECG-gated interventional cardiac vasculature reconstruction

    International Nuclear Information System (INIS)

    Schwemmer, C; Müller, K; Hornegger, J; Rohkohl, C; Lauritsch, G

    2013-01-01

    Three-dimensional reconstruction of cardiac vasculature from angiographic C-arm CT (rotational angiography) data is a major challenge. Motion artefacts corrupt image quality, reducing usability for diagnosis and guidance. Many state-of-the-art approaches depend on retrospective ECG-gating of projection data for image reconstruction. A trade-off has to be made regarding the size of the ECG-gating window. A large temporal window is desirable to avoid undersampling. However, residual motion will occur in a large window, causing motion artefacts. We present an algorithm to correct for residual motion. Our approach is based on a deformable 2D–2D registration between the forward projection of an initial, ECG-gated reconstruction, and the original projection data. The approach is fully automatic and does not require any complex segmentation of vasculature, or landmarks. The estimated motion is compensated for during the backprojection step of a subsequent reconstruction. We evaluated the method using the publicly available CAVAREV platform and on six human clinical datasets. We found a better visibility of structure, reduced motion artefacts, and increased sharpness of the vessels in the compensated reconstructions compared to the initial reconstructions. At the time of writing, our algorithm outperforms the leading result of the CAVAREV ranking list. For the clinical datasets, we found an average reduction of motion artefacts by 13 ± 6%. Vessel sharpness was improved by 25 ± 12% on average. (paper)

  15. Reconstruction of the residual stresses in a hyperelastic body using ultrasound techniques

    KAUST Repository

    Joshi, Sunnie

    2013-09-01

    This paper focuses on a novel approach for characterizing the residual stress field in soft tissue using ultrasound interrogation. A nonlinear inverse spectral technique is developed that makes fundamental use of the finite strain nonlinear response of the material to a quasi-static loading. The soft tissue is modeled as a nonlinear, prestressed and residually stressed, isotropic, slightly compressible elastic body with a rectangular geometry. A boundary value problem is formulated for the residually stressed and prestressed soft tissue, the boundary of which is subjected to a quasi-static pressure, and then an idealized model for the ultrasound interrogation is constructed by superimposing small amplitude time harmonic infinitesimal vibrations on static finite deformation via an asymptotic construction. The model is studied, through a semi-inverse approach, for a specific class of deformations that leads to a system of second order differential equations with homogeneous boundary conditions of Sturm-Liouville type. By making use of the classical theory of inverse Sturm-Liouville problems, and root finding and optimization techniques, several inverse spectral algorithms are developed to approximate the residual stress distribution in the body, given the first few eigenfrequencies of several induced static pressures. © 2013 Elsevier Ltd. All rights reserved.

  16. Fatigue evaluation algorithms: Review

    Energy Technology Data Exchange (ETDEWEB)

    Passipoularidis, V.A.; Broendsted, P.

    2009-11-15

    A progressive damage fatigue simulator for variable amplitude loads named FADAS is discussed in this work. FADAS (Fatigue Damage Simulator) performs ply by ply stress analysis using classical lamination theory and implements adequate stiffness discount tactics based on the failure criterion of Puck, to model the degradation caused by failure events in ply level. Residual strength is incorporated as fatigue damage accumulation metric. Once the typical fatigue and static properties of the constitutive ply are determined,the performance of an arbitrary lay-up under uniaxial and/or multiaxial load time series can be simulated. The predictions are validated against fatigue life data both from repeated block tests at a single stress ratio as well as against spectral fatigue using the WISPER, WISPERX and NEW WISPER load sequences on a Glass/Epoxy multidirectional laminate typical of a wind turbine rotor blade construction. Two versions of the algorithm, the one using single-step and the other using incremental application of each load cycle (in case of ply failure) are implemented and compared. Simulation results confirm the ability of the algorithm to take into account load sequence effects. In general, FADAS performs well in predicting life under both spectral and block loading fatigue. (author)

  17. Improved parallel solution techniques for the integral transport matrix method

    Energy Technology Data Exchange (ETDEWEB)

    Zerr, R. Joseph, E-mail: rjz116@psu.edu [Department of Mechanical and Nuclear Engineering, The Pennsylvania State University, University Park, PA (United States); Azmy, Yousry Y., E-mail: yyazmy@ncsu.edu [Department of Nuclear Engineering, North Carolina State University, Burlington Engineering Laboratories, Raleigh, NC (United States)

    2011-07-01

    Alternative solution strategies to the parallel block Jacobi (PBJ) method for the solution of the global problem with the integral transport matrix method operators have been designed and tested. The most straightforward improvement to the Jacobi iterative method is the Gauss-Seidel alternative. The parallel red-black Gauss-Seidel (PGS) algorithm can improve on the number of iterations and reduce work per iteration by applying an alternating red-black color-set to the subdomains and assigning multiple sub-domains per processor. A parallel GMRES(m) method was implemented as an alternative to stationary iterations. Computational results show that the PGS method can improve on the PBJ method execution time by up to 10´ when eight sub-domains per processor are used. However, compared to traditional source iterations with diffusion synthetic acceleration, it is still approximately an order of magnitude slower. The best-performing cases are optically thick because sub-domains decouple, yielding faster convergence. Further tests revealed that 64 sub-domains per processor was the best performing level of sub-domain division. An acceleration technique that improves the convergence rate would greatly improve the ITMM. The GMRES(m) method with a diagonal block pre conditioner consumes approximately the same time as the PBJ solver but could be improved by an as yet undeveloped, more efficient pre conditioner. (author)

  18. Improved parallel solution techniques for the integral transport matrix method

    International Nuclear Information System (INIS)

    Zerr, R. Joseph; Azmy, Yousry Y.

    2011-01-01

    Alternative solution strategies to the parallel block Jacobi (PBJ) method for the solution of the global problem with the integral transport matrix method operators have been designed and tested. The most straightforward improvement to the Jacobi iterative method is the Gauss-Seidel alternative. The parallel red-black Gauss-Seidel (PGS) algorithm can improve on the number of iterations and reduce work per iteration by applying an alternating red-black color-set to the subdomains and assigning multiple sub-domains per processor. A parallel GMRES(m) method was implemented as an alternative to stationary iterations. Computational results show that the PGS method can improve on the PBJ method execution time by up to 10´ when eight sub-domains per processor are used. However, compared to traditional source iterations with diffusion synthetic acceleration, it is still approximately an order of magnitude slower. The best-performing cases are optically thick because sub-domains decouple, yielding faster convergence. Further tests revealed that 64 sub-domains per processor was the best performing level of sub-domain division. An acceleration technique that improves the convergence rate would greatly improve the ITMM. The GMRES(m) method with a diagonal block pre conditioner consumes approximately the same time as the PBJ solver but could be improved by an as yet undeveloped, more efficient pre conditioner. (author)

  19. Parametric optimization of ultrasonic machining process using gravitational search and fireworks algorithms

    Directory of Open Access Journals (Sweden)

    Debkalpa Goswami

    2015-03-01

    Full Text Available Ultrasonic machining (USM is a mechanical material removal process used to erode holes and cavities in hard or brittle workpieces by using shaped tools, high-frequency mechanical motion and an abrasive slurry. Unlike other non-traditional machining processes, such as laser beam and electrical discharge machining, USM process does not thermally damage the workpiece or introduce significant levels of residual stress, which is important for survival of materials in service. For having enhanced machining performance and better machined job characteristics, it is often required to determine the optimal control parameter settings of an USM process. The earlier mathematical approaches for parametric optimization of USM processes have mostly yielded near optimal or sub-optimal solutions. In this paper, two almost unexplored non-conventional optimization techniques, i.e. gravitational search algorithm (GSA and fireworks algorithm (FWA are applied for parametric optimization of USM processes. The optimization performance of these two algorithms is compared with that of other popular population-based algorithms, and the effects of their algorithm parameters on the derived optimal solutions and computational speed are also investigated. It is observed that FWA provides the best optimal results for the considered USM processes.

  20. A survey of residual analysis and a new test of residual trend.

    Science.gov (United States)

    McDowell, J J; Calvin, Olivia L; Klapes, Bryan

    2016-05-01

    A survey of residual analysis in behavior-analytic research reveals that existing methods are problematic in one way or another. A new test for residual trends is proposed that avoids the problematic features of the existing methods. It entails fitting cubic polynomials to sets of residuals and comparing their effect sizes to those that would be expected if the sets of residuals were random. To this end, sampling distributions of effect sizes for fits of a cubic polynomial to random data were obtained by generating sets of random standardized residuals of various sizes, n. A cubic polynomial was then fitted to each set of residuals and its effect size was calculated. This yielded a sampling distribution of effect sizes for each n. To test for a residual trend in experimental data, the median effect size of cubic-polynomial fits to sets of experimental residuals can be compared to the median of the corresponding sampling distribution of effect sizes for random residuals using a sign test. An example from the literature, which entailed comparing mathematical and computational models of continuous choice, is used to illustrate the utility of the test. © 2016 Society for the Experimental Analysis of Behavior.

  1. Precise Aperture-Dependent Motion Compensation with Frequency Domain Fast Back-Projection Algorithm

    Directory of Open Access Journals (Sweden)

    Man Zhang

    2017-10-01

    Full Text Available Precise azimuth-variant motion compensation (MOCO is an essential and difficult task for high-resolution synthetic aperture radar (SAR imagery. In conventional post-filtering approaches, residual azimuth-variant motion errors are generally compensated through a set of spatial post-filters, where the coarse-focused image is segmented into overlapped blocks concerning the azimuth-dependent residual errors. However, image domain post-filtering approaches, such as precise topography- and aperture-dependent motion compensation algorithm (PTA, have difficulty of robustness in declining, when strong motion errors are involved in the coarse-focused image. In this case, in order to capture the complete motion blurring function within each image block, both the block size and the overlapped part need necessary extension leading to degeneration of efficiency and robustness inevitably. Herein, a frequency domain fast back-projection algorithm (FDFBPA is introduced to deal with strong azimuth-variant motion errors. FDFBPA disposes of the azimuth-variant motion errors based on a precise azimuth spectrum expression in the azimuth wavenumber domain. First, a wavenumber domain sub-aperture processing strategy is introduced to accelerate computation. After that, the azimuth wavenumber spectrum is partitioned into a set of wavenumber blocks, and each block is formed into a sub-aperture coarse resolution image via the back-projection integral. Then, the sub-aperture images are straightforwardly fused together in azimuth wavenumber domain to obtain a full resolution image. Moreover, chirp-Z transform (CZT is also introduced to implement the sub-aperture back-projection integral, increasing the efficiency of the algorithm. By disusing the image domain post-filtering strategy, robustness of the proposed algorithm is improved. Both simulation and real-measured data experiments demonstrate the effectiveness and superiority of the proposal.

  2. Near-Field Three-Dimensional Planar Millimeter-Wave Holographic Imaging by Using Frequency Scaling Algorithm

    Directory of Open Access Journals (Sweden)

    Ye Zhang

    2017-10-01

    Full Text Available In this paper, a fast three-dimensional (3-D frequency scaling algorithm (FSA with large depth of focus is presented for near-field planar millimeter-wave (MMW holographic imaging. Considering the cross-range range coupling term which is neglected in the conventional range migration algorithm (RMA, we propose an algorithm performing the range cell migration correction for de-chirped signals without interpolation by using a 3-D frequency scaling operation. First, to deal with the cross-range range coupling term, a 3-D frequency scaling operator is derived to eliminate the space variation of range cell migration. Then, a range migration correction factor is performed to compensate for the residual range cell migration. Finally, the imaging results are obtained by matched filtering in the cross-range direction. Compared with the conventional RMA, the proposed algorithm is comparable in accuracy but more efficient by using only chirp multiplications and fast Fourier transforms (FFTs. The algorithm has been tested with satisfying results by both simulation and experiment.

  3. Optimisation of centroiding algorithms for photon event counting imaging

    International Nuclear Information System (INIS)

    Suhling, K.; Airey, R.W.; Morgan, B.L.

    1999-01-01

    Approaches to photon event counting imaging in which the output events of an image intensifier are located using a centroiding technique have long been plagued by fixed pattern noise in which a grid of dimensions similar to those of the CCD pixels is superimposed on the image. This is caused by a mismatch between the photon event shape and the centroiding algorithm. We have used hyperbolic cosine, Gaussian, Lorentzian, parabolic as well as 3-, 5-, and 7-point centre of gravity algorithms, and hybrids thereof, to assess means of minimising this fixed pattern noise. We show that fixed pattern noise generated by the widely used centre of gravity centroiding is due to intrinsic features of the algorithm. Our results confirm that the recently proposed use of Gaussian centroiding does indeed show a significant reduction of fixed pattern noise compared to centre of gravity centroiding (Michel et al., Mon. Not. R. Astron. Soc. 292 (1997) 611-620). However, the disadvantage of a Gaussian algorithm is a centroiding failure for small pulses, caused by a division by zero, which leads to a loss of detective quantum efficiency (DQE) and to small amounts of residual fixed pattern noise. Using both real data from an image intensifier system employing a progressive scan camera, framegrabber and PC, and also synthetic data from Monte-Carlo simulations, we find that hybrid centroiding algorithms can reduce the fixed pattern noise without loss of resolution or loss of DQE. Imaging a test pattern to assess the features of the different algorithms shows that a hybrid of Gaussian and 3-point centre of gravity centroiding algorithms results in an optimum combination of low fixed pattern noise (lower than a simple Gaussian), high DQE, and high resolution. The Lorentzian algorithm gives the worst results in terms of high fixed pattern noise and low resolution, and the Gaussian and hyperbolic cosine algorithms have the lowest DQEs

  4. On the Relations between the Attacks on Symmetric Homomorphic Encryption over the Residue Ring

    Directory of Open Access Journals (Sweden)

    Alina V. Trepacheva

    2017-06-01

    Full Text Available The paper considers the security of symmetric homomorphic cryptosystems (HC over the residue ring. The main task is to establish an equivalence between ciphertexts only attack (COA and known plaintexts attack (KPA for HC. The notion of reducibility between attacks and sufficient condition of reducibility from COA to KPA are given for this purpose. The main idea is: to prove reducibility from COA to KPA we need to find a function over residue ring being efficiently computable and having a small image size comparing with the size of residue ring. The study of reducibility existence is important since it allows to understand better the security level of symmetric HC proposed in literature. A vulnerability against KPA has been already found for the majority of these HC. Thus the reducibility presence can demonstrate that cryptosystems under the study are not secure even against COA, and therefore they are totally insecure and shouldn’t be used in practice. We give an example of reducibility from COA to KPA for residue ring being a simple field. Based on this example we show an efficient COA on one symmetric HC for small field. Also we separately consider the case of residue ring composed using number n being hard-to-factor. For such n an efficient algorithm to construct an efficiently computable function with small image is unknown so far. So further work related to cryptanalysis of existing symmetric HC will be directed into study of functions properties over residue rings modulo numbers hard for factorization.

  5. Evaluation of residue-residue contact predictions in CASP9

    KAUST Repository

    Monastyrskyy, Bohdan; Fidelis, Krzysztof; Tramontano, Anna; Kryshtafovych, Andriy

    2011-01-01

    This work presents the results of the assessment of the intramolecular residue-residue contact predictions submitted to CASP9. The methodology for the assessment does not differ from that used in previous CASPs, with two basic evaluation measures

  6. Parameter Estimation of Damped Compound Pendulum Differential Evolution Algorithm

    Directory of Open Access Journals (Sweden)

    Saad Mohd Sazli

    2016-01-01

    Full Text Available This paper present the parameter identification of damped compound pendulum using differential evolution algorithm. The procedure used to achieve the parameter identification of the experimental system consisted of input output data collection, ARX model order selection and parameter estimation using conventional method least square (LS and differential evolution (DE algorithm. PRBS signal is used to be input signal to regulate the motor speed. Whereas, the output signal is taken from position sensor. Both, input and output data is used to estimate the parameter of the ARX model. The residual error between the actual and predicted output responses of the models is validated using mean squares error (MSE. Analysis showed that, MSE value for LS is 0.0026 and MSE value for DE is 3.6601×10-5. Based results obtained, it was found that DE have lower MSE than the LS method.

  7. A multi-sensor burned area algorithm for crop residue burning in northwestern India: validation and sources of error

    Science.gov (United States)

    Liu, T.; Marlier, M. E.; Karambelas, A. N.; Jain, M.; DeFries, R. S.

    2017-12-01

    A leading source of outdoor emissions in northwestern India comes from crop residue burning after the annual monsoon (kharif) and winter (rabi) crop harvests. Agricultural burned area, from which agricultural fire emissions are often derived, can be poorly quantified due to the mismatch between moderate-resolution satellite sensors and the relatively small size and short burn period of the fires. Many previous studies use the Global Fire Emissions Database (GFED), which is based on the Moderate Resolution Imaging Spectroradiometer (MODIS) burned area product MCD64A1, as an outdoor fires emissions dataset. Correction factors with MODIS active fire detections have previously attempted to account for small fires. We present a new burned area classification algorithm that leverages more frequent MODIS observations (500 m x 500 m) with higher spatial resolution Landsat (30 m x 30 m) observations. Our approach is based on two-tailed Normalized Burn Ratio (NBR) thresholds, abbreviated as ModL2T NBR, and results in an estimated 104 ± 55% higher burned area than GFEDv4.1s (version 4, MCD64A1 + small fires correction) in northwestern India during the 2003-2014 winter (October to November) burning seasons. Regional transport of winter fire emissions affect approximately 63 million people downwind. The general increase in burned area (+37% from 2003-2007 to 2008-2014) over the study period also correlates with increased mechanization (+58% in combine harvester usage from 2001-2002 to 2011-2012). Further, we find strong correlation between ModL2T NBR-derived burned area and results of an independent survey (r = 0.68) and previous studies (r = 0.92). Sources of error arise from small median landholding sizes (1-3 ha), heterogeneous spatial distribution of two dominant burning practices (partial and whole field), coarse spatio-temporal satellite resolution, cloud and haze cover, and limited Landsat scene availability. The burned area estimates of this study can be used to build

  8. Algebraic dynamics algorithm: Numerical comparison with Runge-Kutta algorithm and symplectic geometric algorithm

    Institute of Scientific and Technical Information of China (English)

    WANG ShunJin; ZHANG Hua

    2007-01-01

    Based on the exact analytical solution of ordinary differential equations,a truncation of the Taylor series of the exact solution to the Nth order leads to the Nth order algebraic dynamics algorithm.A detailed numerical comparison is presented with Runge-Kutta algorithm and symplectic geometric algorithm for 12 test models.The results show that the algebraic dynamics algorithm can better preserve both geometrical and dynamical fidelity of a dynamical system at a controllable precision,and it can solve the problem of algorithm-induced dissipation for the Runge-Kutta algorithm and the problem of algorithm-induced phase shift for the symplectic geometric algorithm.

  9. Algebraic dynamics algorithm:Numerical comparison with Runge-Kutta algorithm and symplectic geometric algorithm

    Institute of Scientific and Technical Information of China (English)

    2007-01-01

    Based on the exact analytical solution of ordinary differential equations, a truncation of the Taylor series of the exact solution to the Nth order leads to the Nth order algebraic dynamics algorithm. A detailed numerical comparison is presented with Runge-Kutta algorithm and symplectic geometric algorithm for 12 test models. The results show that the algebraic dynamics algorithm can better preserve both geometrical and dynamical fidelity of a dynamical system at a controllable precision, and it can solve the problem of algorithm-induced dissipation for the Runge-Kutta algorithm and the problem of algorithm-induced phase shift for the symplectic geometric algorithm.

  10. The Algorithm for Algorithms: An Evolutionary Algorithm Based on Automatic Designing of Genetic Operators

    Directory of Open Access Journals (Sweden)

    Dazhi Jiang

    2015-01-01

    Full Text Available At present there is a wide range of evolutionary algorithms available to researchers and practitioners. Despite the great diversity of these algorithms, virtually all of the algorithms share one feature: they have been manually designed. A fundamental question is “are there any algorithms that can design evolutionary algorithms automatically?” A more complete definition of the question is “can computer construct an algorithm which will generate algorithms according to the requirement of a problem?” In this paper, a novel evolutionary algorithm based on automatic designing of genetic operators is presented to address these questions. The resulting algorithm not only explores solutions in the problem space like most traditional evolutionary algorithms do, but also automatically generates genetic operators in the operator space. In order to verify the performance of the proposed algorithm, comprehensive experiments on 23 well-known benchmark optimization problems are conducted. The results show that the proposed algorithm can outperform standard differential evolution algorithm in terms of convergence speed and solution accuracy which shows that the algorithm designed automatically by computers can compete with the algorithms designed by human beings.

  11. An ensemble based nonlinear orthogonal matching pursuit algorithm for sparse history matching of reservoir models

    KAUST Repository

    Fsheikh, Ahmed H.

    2013-01-01

    A nonlinear orthogonal matching pursuit (NOMP) for sparse calibration of reservoir models is presented. Sparse calibration is a challenging problem as the unknowns are both the non-zero components of the solution and their associated weights. NOMP is a greedy algorithm that discovers at each iteration the most correlated components of the basis functions with the residual. The discovered basis (aka support) is augmented across the nonlinear iterations. Once the basis functions are selected from the dictionary, the solution is obtained by applying Tikhonov regularization. The proposed algorithm relies on approximate gradient estimation using an iterative stochastic ensemble method (ISEM). ISEM utilizes an ensemble of directional derivatives to efficiently approximate gradients. In the current study, the search space is parameterized using an overcomplete dictionary of basis functions built using the K-SVD algorithm.

  12. Rigid Residue Scan Simulations Systematically Reveal Residue Entropic Roles in Protein Allostery.

    Directory of Open Access Journals (Sweden)

    Robert Kalescky

    2016-04-01

    Full Text Available Intra-protein information is transmitted over distances via allosteric processes. This ubiquitous protein process allows for protein function changes due to ligand binding events. Understanding protein allostery is essential to understanding protein functions. In this study, allostery in the second PDZ domain (PDZ2 in the human PTP1E protein is examined as model system to advance a recently developed rigid residue scan method combining with configurational entropy calculation and principal component analysis. The contributions from individual residues to whole-protein dynamics and allostery were systematically assessed via rigid body simulations of both unbound and ligand-bound states of the protein. The entropic contributions of individual residues to whole-protein dynamics were evaluated based on covariance-based correlation analysis of all simulations. The changes of overall protein entropy when individual residues being held rigid support that the rigidity/flexibility equilibrium in protein structure is governed by the La Châtelier's principle of chemical equilibrium. Key residues of PDZ2 allostery were identified with good agreement with NMR studies of the same protein bound to the same peptide. On the other hand, the change of entropic contribution from each residue upon perturbation revealed intrinsic differences among all the residues. The quasi-harmonic and principal component analyses of simulations without rigid residue perturbation showed a coherent allosteric mode from unbound and bound states, respectively. The projection of simulations with rigid residue perturbation onto coherent allosteric modes demonstrated the intrinsic shifting of ensemble distributions supporting the population-shift theory of protein allostery. Overall, the study presented here provides a robust and systematic approach to estimate the contribution of individual residue internal motion to overall protein dynamics and allostery.

  13. Lifetime and residual strength of wood subjected to static and variable load

    DEFF Research Database (Denmark)

    Nielsen, Lauge Fuglsang

    1997-01-01

    of load amplitude, load average, fractional time under maximum load, and load frequency. The analysis includes prediction of residual strength (re-cycle strength) during the process of load cycling. It is concluded that number of cycles to failure is a very poor design criterion. The theory......).It is demonstrated how the theory developed can be generalised also to consider non-harmonic load variations. An algorithm is presented for this purpose which might be suggested as a qualified alternative to the Palmgren-Miner’s method normally used in fatigue analysis of materials under arbitrary load variations...

  14. Ensemble Kalman Filtering with Residual Nudging: An Extension to State Estimation Problems with Nonlinear Observation Operators

    KAUST Repository

    Luo, Xiaodong

    2014-10-01

    The ensemble Kalman filter (EnKF) is an efficient algorithm for many data assimilation problems. In certain circumstances, however, divergence of the EnKF might be spotted. In previous studies, the authors proposed an observation-space-based strategy, called residual nudging, to improve the stability of the EnKF when dealing with linear observation operators. The main idea behind residual nudging is to monitor and, if necessary, adjust the distances (misfits) between the real observations and the simulated ones of the state estimates, in the hope that by doing so one may be able to obtain better estimation accuracy. In the present study, residual nudging is extended and modified in order to handle nonlinear observation operators. Such extension and modification result in an iterative filtering framework that, under suitable conditions, is able to achieve the objective of residual nudging for data assimilation problems with nonlinear observation operators. The 40-dimensional Lorenz-96 model is used to illustrate the performance of the iterative filter. Numerical results show that, while a normal EnKF may diverge with nonlinear observation operators, the proposed iterative filter remains stable and leads to reasonable estimation accuracy under various experimental settings.

  15. A comparison of three-dimensional nonequilibrium solution algorithms applied to hypersonic flows with stiff chemical source terms

    Science.gov (United States)

    Palmer, Grant; Venkatapathy, Ethiraj

    1993-01-01

    Three solution algorithms, explicit underrelaxation, point implicit, and lower upper symmetric Gauss-Seidel (LUSGS), are used to compute nonequilibrium flow around the Apollo 4 return capsule at 62 km altitude. By varying the Mach number, the efficiency and robustness of the solution algorithms were tested for different levels of chemical stiffness. The performance of the solution algorithms degraded as the Mach number and stiffness of the flow increased. At Mach 15, 23, and 30, the LUSGS method produces an eight order of magnitude drop in the L2 norm of the energy residual in 1/3 to 1/2 the Cray C-90 computer time as compared to the point implicit and explicit under-relaxation methods. The explicit under-relaxation algorithm experienced convergence difficulties at Mach 23 and above. At Mach 40 the performance of the LUSGS algorithm deteriorates to the point it is out-performed by the point implicit method. The effects of the viscous terms are investigated. Grid dependency questions are explored.

  16. A novel remote sensing algorithm to quantify phycocyanin in cyanobacterial algal blooms

    International Nuclear Information System (INIS)

    Mishra, S; Mishra, D R

    2014-01-01

    We present a novel three-band algorithm (PC 3 ) to retrieve phycocyanin (PC) pigment concentration in cyanobacteria laden inland waters. The water sample and remote sensing reflectance data used for PC 3 calibration and validation were acquired from highly turbid productive catfish aquaculture ponds. Since the characteristic PC absorption feature at 620 nm is contaminated with residual chlorophyll-a (Chl-a) absorption, we propose a coefficient (ψ) for isolating the PC absorption component at 620 nm. Results show that inclusion of the model coefficient relating Chl-a absorption at 620 nm–665 nm enables PC 3 to compensate for the confounding effect of Chl-a at the PC absorption band and considerably increases the accuracy of the PC prediction algorithm. In the current dataset, PC 3 produced the lowest mean relative error of prediction among all PC algorithms considered in this research. Moreover, PC 3 eliminates the nonlinear sensitivity issue of PC algorithms particularly at high PC range (>100 μg L −1 ). Therefore, introduction of PC 3 will have an immediate positive impact on studies monitoring inland and coastal cyanobacterial harmful algal blooms. (letter)

  17. Analysis of the unstressed lattice spacing, d0, for the determination of the residual stress in a friction stir welded plate of an age-hardenable aluminum alloy – Use of equilibrium conditions and a genetic algorithm

    International Nuclear Information System (INIS)

    Cioffi, F.; Hidalgo, J.I.; Fernández, R.; Pirling, T.; Fernández, B.; Gesto, D.; Puente Orench, I.; Rey, P.; González-Doncel, G.

    2014-01-01

    Procedures based on equilibrium conditions (stress and bending moment) have been used to obtain an unstressed lattice spacing, d 0 , as a crucial requirement for calculating the residual stress (RS) profile across a joint conducted on a 10 mm thick plate of age-hardenable AA2024 alloy by friction stir welding (FSW). Two procedures have been used that take advantage of neutron diffraction measurements. First, equilibrium conditions were imposed on sections parallel to the weld so that a constant d 0 value corresponding to the base material region could be calculated analytically. Second, balance conditions were imposed on a section transverse to the weld. Then, using the data and a genetic algorithm, suitable d 0 values for the different regions of the weld have been found. For several reasons, the comb method has proved to be inappropriate for RS determination in the case of age-hardenable alloys. However, the equilibrium conditions, together with the genetic algorithm, has been shown to be very suitable for determining RS profiles in FSW joints of these alloys, where inherent microstructural variations of d 0 across the weld are expected

  18. Newton-like methods for Navier-Stokes solution

    Science.gov (United States)

    Qin, N.; Xu, X.; Richards, B. E.

    1992-12-01

    The paper reports on Newton-like methods called SFDN-alpha-GMRES and SQN-alpha-GMRES methods that have been devised and proven as powerful schemes for large nonlinear problems typical of viscous compressible Navier-Stokes solutions. They can be applied using a partially converged solution from a conventional explicit or approximate implicit method. Developments have included the efficient parallelization of the schemes on a distributed memory parallel computer. The methods are illustrated using a RISC workstation and a transputer parallel system respectively to solve a hypersonic vortical flow.

  19. Triangular preconditioners for saddle point problems with a penalty term

    Energy Technology Data Exchange (ETDEWEB)

    Klawonn, A. [Westfaelische Wilhelms-Universitaet, Muenster (Germany)

    1996-12-31

    Triangular preconditioners for a class of saddle point problems with a penalty term are considered. An important example is the mixed formulation of the pure displacement problem in linear elasticity. It is shown that the spectrum of the preconditioned system is contained in a real, positive interval, and that the interval bounds can be made independent of the discretization and penalty parameters. This fact is used to construct bounds of the convergence rate of the GMRES method used with an energy norm. Numerical results are given for GMRES and BI-CGSTAB.

  20. Residual gas analysis

    International Nuclear Information System (INIS)

    Berecz, I.

    1982-01-01

    Determination of the residual gas composition in vacuum systems by a special mass spectrometric method was presented. The quadrupole mass spectrometer (QMS) and its application in thin film technology was discussed. Results, partial pressure versus time curves as well as the line spectra of the residual gases in case of the vaporization of a Ti-Pd-Au alloy were demonstrated together with the possible construction schemes of QMS residual gas analysers. (Sz.J.)

  1. Assessing the Availability of Wood Residues and Residue Markets in Virginia

    OpenAIRE

    Alderman, Delton R. Jr.

    1998-01-01

    A statewide mail survey of primary and secondary wood product manufacturers was undertaken to quantify the production and consumption of wood residues in Virginia. Two hundred and sixty-six wood product manufacturers responded to the study and they provided information on the production, consumption, markets, income or disposal costs, and disposal methods of wood residues. Hardwood and pine sawmills produce approximately 66 percent of Virginia's wood residues. Virginia's wood product man...

  2. A novel algorithm for a precise analysis of subchondral bone alterations

    Science.gov (United States)

    Gao, Liang; Orth, Patrick; Goebel, Lars K. H.; Cucchiarini, Magali; Madry, Henning

    2016-01-01

    Subchondral bone alterations are emerging as considerable clinical problems associated with articular cartilage repair. Their analysis exposes a pattern of variable changes, including intra-lesional osteophytes, residual microfracture holes, peri-hole bone resorption, and subchondral bone cysts. A precise distinction between them is becoming increasingly important. Here, we present a tailored algorithm based on continuous data to analyse subchondral bone changes using micro-CT images, allowing for a clear definition of each entity. We evaluated this algorithm using data sets originating from two large animal models of osteochondral repair. Intra-lesional osteophytes were detected in 3 of 10 defects in the minipig and in 4 of 5 defects in the sheep model. Peri-hole bone resorption was found in 22 of 30 microfracture holes in the minipig and in 17 of 30 microfracture holes in the sheep model. Subchondral bone cysts appeared in 1 microfracture hole in the minipig and in 5 microfracture holes in the sheep model (n = 30 holes each). Calculation of inter-rater agreement (90% agreement) and Cohen’s kappa (kappa = 0.874) revealed that the novel algorithm is highly reliable, reproducible, and valid. Comparison analysis with the best existing semi-quantitative evaluation method was also performed, supporting the enhanced precision of this algorithm. PMID:27596562

  3. Mutational properties of amino acid residues: implications for evolvability of phosphorylatable residues

    DEFF Research Database (Denmark)

    Creixell, Pau; Schoof, Erwin M.; Tan, Chris Soon Heng

    2012-01-01

    in terms of their mutational activity. Moreover, we highlight the importance of the genetic code and physico-chemical properties of the amino acid residues as likely causes of these inequalities and uncover serine as a mutational hot spot. Finally, we explore the consequences that these different......; it is typically assumed that all amino acid residues are equally likely to mutate or to result from a mutation. Here, by reconstructing ancestral sequences and computing mutational probabilities for all the amino acid residues, we refute this assumption and show extensive inequalities between different residues...... mutational properties have on phosphorylation site evolution, showing that a higher degree of evolvability exists for phosphorylated threonine and, to a lesser extent, serine in comparison with tyrosine residues. As exemplified by the suppression of serine's mutational activity in phosphorylation sites, our...

  4. Developing Benthic Class Specific, Chlorophyll-a Retrieving Algorithms for Optically-Shallow Water Using SeaWiFS

    Directory of Open Access Journals (Sweden)

    Tara Blakey

    2016-10-01

    Full Text Available This study evaluated the ability to improve Sea-Viewing Wide Field-of-View Sensor (SeaWiFS chl-a retrieval from optically shallow coastal waters by applying algorithms specific to the pixels’ benthic class. The form of the Ocean Color (OC algorithm was assumed for this study. The operational atmospheric correction producing Level 2 SeaWiFS data was retained since the focus of this study was on establishing the benefit from the alternative specification of the bio-optical algorithm. Benthic class was determined through satellite image-based classification methods. Accuracy of the chl-a algorithms evaluated was determined through comparison with coincident in situ measurements of chl-a. The regionally-tuned models that were allowed to vary by benthic class produced more accurate estimates of chl-a than the single, unified regionally-tuned model. Mean absolute percent difference was approximately 70% for the regionally-tuned, benthic class-specific algorithms. Evaluation of the residuals indicated the potential for further improvement to chl-a estimation through finer characterization of benthic environments. Atmospheric correction procedures specialized to coastal environments were recognized as areas for future improvement as these procedures would improve both classification and algorithm tuning.

  5. Identification of coevolving residues and coevolution potentials emphasizing structure, bond formation and catalytic coordination in protein evolution.

    Directory of Open Access Journals (Sweden)

    Daniel Y Little

    Full Text Available The structure and function of a protein is dependent on coordinated interactions between its residues. The selective pressures associated with a mutation at one site should therefore depend on the amino acid identity of interacting sites. Mutual information has previously been applied to multiple sequence alignments as a means of detecting coevolutionary interactions. Here, we introduce a refinement of the mutual information method that: 1 removes a significant, non-coevolutionary bias and 2 accounts for heteroscedasticity. Using a large, non-overlapping database of protein alignments, we demonstrate that predicted coevolving residue-pairs tend to lie in close physical proximity. We introduce coevolution potentials as a novel measure of the propensity for the 20 amino acids to pair amongst predicted coevolutionary interactions. Ionic, hydrogen, and disulfide bond-forming pairs exhibited the highest potentials. Finally, we demonstrate that pairs of catalytic residues have a significantly increased likelihood to be identified as coevolving. These correlations to distinct protein features verify the accuracy of our algorithm and are consistent with a model of coevolution in which selective pressures towards preserving residue interactions act to shape the mutational landscape of a protein by restricting the set of admissible neutral mutations.

  6. Evaluation of residue-residue contact prediction in CASP10

    KAUST Repository

    Monastyrskyy, Bohdan

    2013-08-31

    We present the results of the assessment of the intramolecular residue-residue contact predictions from 26 prediction groups participating in the 10th round of the CASP experiment. The most recently developed direct coupling analysis methods did not take part in the experiment likely because they require a very deep sequence alignment not available for any of the 114 CASP10 targets. The performance of contact prediction methods was evaluated with the measures used in previous CASPs (i.e., prediction accuracy and the difference between the distribution of the predicted contacts and that of all pairs of residues in the target protein), as well as new measures, such as the Matthews correlation coefficient, the area under the precision-recall curve and the ranks of the first correctly and incorrectly predicted contact. We also evaluated the ability to detect interdomain contacts and tested whether the difficulty of predicting contacts depends upon the protein length and the depth of the family sequence alignment. The analyses were carried out on the target domains for which structural homologs did not exist or were difficult to identify. The evaluation was performed for all types of contacts (short, medium, and long-range), with emphasis placed on long-range contacts, i.e. those involving residues separated by at least 24 residues along the sequence. The assessment suggests that the best CASP10 contact prediction methods perform at approximately the same level, and comparably to those participating in CASP9.

  7. Proceedings of the 7. biennial residue-to-revenue residual wood conference 2007

    International Nuclear Information System (INIS)

    Raulin, J.

    2007-01-01

    This conference provided information on the highest and best use of residual wood, which is quickly becoming a valuable commodity. Issues concerning forest residues, sawmill wastes, agricultural residues and urban organic materials were discussed along with trends in Canadian surplus mill waste production. The evolving nature and technologies of the biomass business were highlighted with particular focus on how to generate energy and save money through the use of residual wood. Residual wood energy projects and developments in Canada, North America and Europe were outlined along with biomass development in relation to forest fires and insect disturbances. Cogeneration technologies using wood wastes for thermal heat, steam and electricity were also presented, along with transportation fuel technologies for the production of ethanol. It was noted that with the rising cost of energy, the forest industry is seeking energy solutions based on the use of residual wood. The range of economically practical residual wood solutions continues to grow as energy prices increase. The conference was attended by more than 200 delegates from the forest industry, suppliers and government representatives, to discuss policies and procedures currently in place. Industry investment is being stimulated by the potential for biofuels and biochemicals, as well as the co-operation between the forest and energy sectors. The conference featured 23 presentations, of which 12 have been catalogued separately for inclusion in this database. refs., tabs., figs

  8. ALFA: an automated line fitting algorithm

    Science.gov (United States)

    Wesson, R.

    2016-03-01

    I present the automated line fitting algorithm, ALFA, a new code which can fit emission line spectra of arbitrary wavelength coverage and resolution, fully automatically. In contrast to traditional emission line fitting methods which require the identification of spectral features suspected to be emission lines, ALFA instead uses a list of lines which are expected to be present to construct a synthetic spectrum. The parameters used to construct the synthetic spectrum are optimized by means of a genetic algorithm. Uncertainties are estimated using the noise structure of the residuals. An emission line spectrum containing several hundred lines can be fitted in a few seconds using a single processor of a typical contemporary desktop or laptop PC. I show that the results are in excellent agreement with those measured manually for a number of spectra. Where discrepancies exist, the manually measured fluxes are found to be less accurate than those returned by ALFA. Together with the code NEAT, ALFA provides a powerful way to rapidly extract physical information from observations, an increasingly vital function in the era of highly multiplexed spectroscopy. The two codes can deliver a reliable and comprehensive analysis of very large data sets in a few hours with little or no user interaction.

  9. Solid residues

    International Nuclear Information System (INIS)

    Mulder, E.; Duin, P.J. van; Grootenboer, G.J.

    1995-01-01

    A summary is presented of the many investigations that have been done on solid residues of atmospheric fluid bed combustion (AFBC). These residues are bed ash, cyclone ash and bag filter ash. Physical and chemical properties are discussed and then the various uses of residues (in fillers, bricks, gravel, and for recovery of aluminium) are summarised. Toxicological properties of fly ash and stack ash are discussed as are risks of pneumoconiosis for workers handling fly ash, and contamination of water by ashes. On the basis of present information it is concluded that risks to public health from exposure to emissions of coal fly ash from AFBC appear small or negligible as are health risk to workers in the coal fly ash processing industry. 35 refs., 5 figs., 12 tabs

  10. A Minimum Variance Algorithm for Overdetermined TOA Equations with an Altitude Constraint.

    Energy Technology Data Exchange (ETDEWEB)

    Romero, Louis A; Mason, John J.

    2018-04-01

    We present a direct (non-iterative) method for solving for the location of a radio frequency (RF) emitter, or an RF navigation receiver, using four or more time of arrival (TOA) measurements and an assumed altitude above an ellipsoidal earth. Both the emitter tracking problem and the navigation application are governed by the same equations, but with slightly different interpreta- tions of several variables. We treat the assumed altitude as a soft constraint, with a specified noise level, just as the TOA measurements are handled, with their respective noise levels. With 4 or more TOA measurements and the assumed altitude, the problem is overdetermined and is solved in the weighted least squares sense for the 4 unknowns, the 3-dimensional position and time. We call the new technique the TAQMV (TOA Altitude Quartic Minimum Variance) algorithm, and it achieves the minimum possible error variance for given levels of TOA and altitude estimate noise. The method algebraically produces four solutions, the least-squares solution, and potentially three other low residual solutions, if they exist. In the lightly overdermined cases where multiple local minima in the residual error surface are more likely to occur, this algebraic approach can produce all of the minima even when an iterative approach fails to converge. Algorithm performance in terms of solution error variance and divergence rate for bas eline (iterative) and proposed approach are given in tables.

  11. Structure optimization by heuristic algorithm in a coarse-grained off-lattice model

    International Nuclear Information System (INIS)

    Jing-Fa, Liu

    2009-01-01

    A heuristic algorithm is presented for a three-dimensional off-lattice AB model consisting of hydrophobic (A) and hydrophilic (B) residues in Fibonacci sequences. By incorporating extra energy contributions into the original potential function, we convert the constrained optimization problem of AB model into an unconstrained optimization problem which can be solved by the gradient method. After the gradient minimization leads to the basins of the local energy minima, the heuristic off-trap strategy and subsequent neighborhood search mechanism are then proposed to get out of local minima and search for the lower-energy configurations. Furthermore, in order to improve the efficiency of the proposed algorithm, we apply the improved version called the new PERM with importance sampling (nPERMis) of the chain-growth algorithm, pruned-enriched-Rosenbluth method (PERM), to face-centered-cubic (FCC)-lattice to produce the initial configurations. The numerical results show that the proposed methods are very promising for finding the ground states of proteins. In several cases, we found the ground state energies are lower than the best values reported in the present literature

  12. A Novel Online Data-Driven Algorithm for Detecting UAV Navigation Sensor Faults.

    Science.gov (United States)

    Sun, Rui; Cheng, Qi; Wang, Guanyu; Ochieng, Washington Yotto

    2017-09-29

    The use of Unmanned Aerial Vehicles (UAVs) has increased significantly in recent years. On-board integrated navigation sensors are a key component of UAVs' flight control systems and are essential for flight safety. In order to ensure flight safety, timely and effective navigation sensor fault detection capability is required. In this paper, a novel data-driven Adaptive Neuron Fuzzy Inference System (ANFIS)-based approach is presented for the detection of on-board navigation sensor faults in UAVs. Contrary to the classic UAV sensor fault detection algorithms, based on predefined or modelled faults, the proposed algorithm combines an online data training mechanism with the ANFIS-based decision system. The main advantages of this algorithm are that it allows real-time model-free residual analysis from Kalman Filter (KF) estimates and the ANFIS to build a reliable fault detection system. In addition, it allows fast and accurate detection of faults, which makes it suitable for real-time applications. Experimental results have demonstrated the effectiveness of the proposed fault detection method in terms of accuracy and misdetection rate.

  13. A Novel Online Data-Driven Algorithm for Detecting UAV Navigation Sensor Faults

    Directory of Open Access Journals (Sweden)

    Rui Sun

    2017-09-01

    Full Text Available The use of Unmanned Aerial Vehicles (UAVs has increased significantly in recent years. On-board integrated navigation sensors are a key component of UAVs’ flight control systems and are essential for flight safety. In order to ensure flight safety, timely and effective navigation sensor fault detection capability is required. In this paper, a novel data-driven Adaptive Neuron Fuzzy Inference System (ANFIS-based approach is presented for the detection of on-board navigation sensor faults in UAVs. Contrary to the classic UAV sensor fault detection algorithms, based on predefined or modelled faults, the proposed algorithm combines an online data training mechanism with the ANFIS-based decision system. The main advantages of this algorithm are that it allows real-time model-free residual analysis from Kalman Filter (KF estimates and the ANFIS to build a reliable fault detection system. In addition, it allows fast and accurate detection of faults, which makes it suitable for real-time applications. Experimental results have demonstrated the effectiveness of the proposed fault detection method in terms of accuracy and misdetection rate.

  14. Algorithms

    Indian Academy of Sciences (India)

    polynomial) division have been found in Vedic Mathematics which are dated much before Euclid's algorithm. A programming language Is used to describe an algorithm for execution on a computer. An algorithm expressed using a programming.

  15. On the Impact of Localization and Density Control Algorithms in Target Tracking Applications for Wireless Sensor Networks

    Science.gov (United States)

    Campos, Andre N.; Souza, Efren L.; Nakamura, Fabiola G.; Nakamura, Eduardo F.; Rodrigues, Joel J. P. C.

    2012-01-01

    Target tracking is an important application of wireless sensor networks. The networks' ability to locate and track an object is directed linked to the nodes' ability to locate themselves. Consequently, localization systems are essential for target tracking applications. In addition, sensor networks are often deployed in remote or hostile environments. Therefore, density control algorithms are used to increase network lifetime while maintaining its sensing capabilities. In this work, we analyze the impact of localization algorithms (RPE and DPE) and density control algorithms (GAF, A3 and OGDC) on target tracking applications. We adapt the density control algorithms to address the k-coverage problem. In addition, we analyze the impact of network density, residual integration with density control, and k-coverage on both target tracking accuracy and network lifetime. Our results show that DPE is a better choice for target tracking applications than RPE. Moreover, among the evaluated density control algorithms, OGDC is the best option among the three. Although the choice of the density control algorithm has little impact on the tracking precision, OGDC outperforms GAF and A3 in terms of tracking time. PMID:22969329

  16. Handling of Solid Residues

    International Nuclear Information System (INIS)

    Medina Bermudez, Clara Ines

    1999-01-01

    The topic of solid residues is specifically of great interest and concern for the authorities, institutions and community that identify in them a true threat against the human health and the atmosphere in the related with the aesthetic deterioration of the urban centers and of the natural landscape; in the proliferation of vectorial transmitters of illnesses and the effect on the biodiversity. Inside the wide spectrum of topics that they keep relationship with the environmental protection, the inadequate handling of solid residues and residues dangerous squatter an important line in the definition of political and practical environmentally sustainable. The industrial development and the population's growth have originated a continuous increase in the production of solid residues; of equal it forms, their composition day after day is more heterogeneous. The base for the good handling includes the appropriate intervention of the different stages of an integral administration of residues, which include the separation in the source, the gathering, the handling, the use, treatment, final disposition and the institutional organization of the administration. The topic of the dangerous residues generates more expectation. These residues understand from those of pathogen type that are generated in the establishments of health that of hospital attention, until those of combustible, inflammable type, explosive, radio-active, volatile, corrosive, reagent or toxic, associated to numerous industrial processes, common in our countries in development

  17. STAR Algorithm Integration Team - Facilitating operational algorithm development

    Science.gov (United States)

    Mikles, V. J.

    2015-12-01

    The NOAA/NESDIS Center for Satellite Research and Applications (STAR) provides technical support of the Joint Polar Satellite System (JPSS) algorithm development and integration tasks. Utilizing data from the S-NPP satellite, JPSS generates over thirty Environmental Data Records (EDRs) and Intermediate Products (IPs) spanning atmospheric, ocean, cryosphere, and land weather disciplines. The Algorithm Integration Team (AIT) brings technical expertise and support to product algorithms, specifically in testing and validating science algorithms in a pre-operational environment. The AIT verifies that new and updated algorithms function in the development environment, enforces established software development standards, and ensures that delivered packages are functional and complete. AIT facilitates the development of new JPSS-1 algorithms by implementing a review approach based on the Enterprise Product Lifecycle (EPL) process. Building on relationships established during the S-NPP algorithm development process and coordinating directly with science algorithm developers, the AIT has implemented structured reviews with self-contained document suites. The process has supported algorithm improvements for products such as ozone, active fire, vegetation index, and temperature and moisture profiles.

  18. The Nonlocal Sparse Reconstruction Algorithm by Similarity Measurement with Shearlet Feature Vector

    Directory of Open Access Journals (Sweden)

    Wu Qidi

    2014-01-01

    Full Text Available Due to the limited accuracy of conventional methods with image restoration, the paper supplied a nonlocal sparsity reconstruction algorithm with similarity measurement. To improve the performance of restoration results, we proposed two schemes to dictionary learning and sparse coding, respectively. In the part of the dictionary learning, we measured the similarity between patches from degraded image by constructing the Shearlet feature vector. Besides, we classified the patches into different classes with similarity and trained the cluster dictionary for each class, by cascading which we could gain the universal dictionary. In the part of sparse coding, we proposed a novel optimal objective function with the coding residual item, which can suppress the residual between the estimate coding and true sparse coding. Additionally, we show the derivation of self-adaptive regularization parameter in optimization under the Bayesian framework, which can make the performance better. It can be indicated from the experimental results that by taking full advantage of similar local geometric structure feature existing in the nonlocal patches and the coding residual suppression, the proposed method shows advantage both on visual perception and PSNR compared to the conventional methods.

  19. Integrated Fault Diagnosis Algorithm for Motor Sensors of In-Wheel Independent Drive Electric Vehicles

    Science.gov (United States)

    Jeon, Namju; Lee, Hyeongcheol

    2016-01-01

    An integrated fault-diagnosis algorithm for a motor sensor of in-wheel independent drive electric vehicles is presented. This paper proposes a method that integrates the high- and low-level fault diagnoses to improve the robustness and performance of the system. For the high-level fault diagnosis of vehicle dynamics, a planar two-track non-linear model is first selected, and the longitudinal and lateral forces are calculated. To ensure redundancy of the system, correlation between the sensor and residual in the vehicle dynamics is analyzed to detect and separate the fault of the drive motor system of each wheel. To diagnose the motor system for low-level faults, the state equation of an interior permanent magnet synchronous motor is developed, and a parity equation is used to diagnose the fault of the electric current and position sensors. The validity of the high-level fault-diagnosis algorithm is verified using Carsim and Matlab/Simulink co-simulation. The low-level fault diagnosis is verified through Matlab/Simulink simulation and experiments. Finally, according to the residuals of the high- and low-level fault diagnoses, fault-detection flags are defined. On the basis of this information, an integrated fault-diagnosis strategy is proposed. PMID:27973431

  20. Integrated Fault Diagnosis Algorithm for Motor Sensors of In-Wheel Independent Drive Electric Vehicles.

    Science.gov (United States)

    Jeon, Namju; Lee, Hyeongcheol

    2016-12-12

    An integrated fault-diagnosis algorithm for a motor sensor of in-wheel independent drive electric vehicles is presented. This paper proposes a method that integrates the high- and low-level fault diagnoses to improve the robustness and performance of the system. For the high-level fault diagnosis of vehicle dynamics, a planar two-track non-linear model is first selected, and the longitudinal and lateral forces are calculated. To ensure redundancy of the system, correlation between the sensor and residual in the vehicle dynamics is analyzed to detect and separate the fault of the drive motor system of each wheel. To diagnose the motor system for low-level faults, the state equation of an interior permanent magnet synchronous motor is developed, and a parity equation is used to diagnose the fault of the electric current and position sensors. The validity of the high-level fault-diagnosis algorithm is verified using Carsim and Matlab/Simulink co-simulation. The low-level fault diagnosis is verified through Matlab/Simulink simulation and experiments. Finally, according to the residuals of the high- and low-level fault diagnoses, fault-detection flags are defined. On the basis of this information, an integrated fault-diagnosis strategy is proposed.

  1. Structural assessment of aerospace components using image processing algorithms and Finite Element models

    DEFF Research Database (Denmark)

    Stamatelos, Dimtrios; Kappatos, Vassilios

    2017-01-01

    Purpose – This paper presents the development of an advanced structural assessment approach for aerospace components (metallic and composites). This work focuses on developing an automatic image processing methodology based on Non Destructive Testing (NDT) data and numerical models, for predicting...... the residual strength of these components. Design/methodology/approach – An image processing algorithm, based on the threshold method, has been developed to process and quantify the geometric characteristics of damages. Then, a parametric Finite Element (FE) model of the damaged component is developed based...... on the inputs acquired from the image processing algorithm. The analysis of the metallic structures is employing the Extended FE Method (XFEM), while for the composite structures the Cohesive Zone Model (CZM) technique with Progressive Damage Modelling (PDM) is used. Findings – The numerical analyses...

  2. Adjustment Criterion and Algorithm in Adjustment Model with Uncertain

    Directory of Open Access Journals (Sweden)

    SONG Yingchun

    2015-02-01

    Full Text Available Uncertainty often exists in the process of obtaining measurement data, which affects the reliability of parameter estimation. This paper establishes a new adjustment model in which uncertainty is incorporated into the function model as a parameter. A new adjustment criterion and its iterative algorithm are given based on uncertainty propagation law in the residual error, in which the maximum possible uncertainty is minimized. This paper also analyzes, with examples, the different adjustment criteria and features of optimal solutions about the least-squares adjustment, the uncertainty adjustment and total least-squares adjustment. Existing error theory is extended with new observational data processing method about uncertainty.

  3. A multiresolution hierarchical classification algorithm for filtering airborne LiDAR data

    Science.gov (United States)

    Chen, Chuanfa; Li, Yanyan; Li, Wei; Dai, Honglei

    2013-08-01

    We presented a multiresolution hierarchical classification (MHC) algorithm for differentiating ground from non-ground LiDAR point cloud based on point residuals from the interpolated raster surface. MHC includes three levels of hierarchy, with the simultaneous increase of cell resolution and residual threshold from the low to the high level of the hierarchy. At each level, the surface is iteratively interpolated towards the ground using thin plate spline (TPS) until no ground points are classified, and the classified ground points are used to update the surface in the next iteration. 15 groups of benchmark dataset, provided by the International Society for Photogrammetry and Remote Sensing (ISPRS) commission, were used to compare the performance of MHC with those of the 17 other publicized filtering methods. Results indicated that MHC with the average total error and average Cohen’s kappa coefficient of 4.11% and 86.27% performs better than all other filtering methods.

  4. Selfish Gene Algorithm Vs Genetic Algorithm: A Review

    Science.gov (United States)

    Ariff, Norharyati Md; Khalid, Noor Elaiza Abdul; Hashim, Rathiah; Noor, Noorhayati Mohamed

    2016-11-01

    Evolutionary algorithm is one of the algorithms inspired by the nature. Within little more than a decade hundreds of papers have reported successful applications of EAs. In this paper, the Selfish Gene Algorithms (SFGA), as one of the latest evolutionary algorithms (EAs) inspired from the Selfish Gene Theory which is an interpretation of Darwinian Theory ideas from the biologist Richards Dawkins on 1989. In this paper, following a brief introduction to the Selfish Gene Algorithm (SFGA), the chronology of its evolution is presented. It is the purpose of this paper is to present an overview of the concepts of Selfish Gene Algorithm (SFGA) as well as its opportunities and challenges. Accordingly, the history, step involves in the algorithm are discussed and its different applications together with an analysis of these applications are evaluated.

  5. A noise-resistant ADSA-PH algorithm for superhydrophobic surface’s static contact angle evaluation

    OpenAIRE

    Z. N. Xu

    2017-01-01

    The blur around the contact points significantly decreases the evaluated static contact angle for superhydrophobic surface which is clearly presented in the paper. To improve the accuracy in the evaluated static contact angle for superhydrophobic surface, an accurate static contact angle algorithm, namely ADSA-PH (axisymmetric drop shape analysis-profile and height), is proposed. It discards the extracted drop edge points close to the contact points and makes use of the residual points and th...

  6. Algorithms

    Indian Academy of Sciences (India)

    to as 'divide-and-conquer'. Although there has been a large effort in realizing efficient algorithms, there are not many universally accepted algorithm design paradigms. In this article, we illustrate algorithm design techniques such as balancing, greedy strategy, dynamic programming strategy, and backtracking or traversal of ...

  7. PENETAPAN RESIDU DAN PERKIRAAN PENETAPAN BATAS MAKSIMUM RESIDU (BMR ORGANOKLORIN PADA SIMPLISIA

    Directory of Open Access Journals (Sweden)

    Ani Isnawati

    2012-10-01

    Full Text Available Penggunaaan bahan obat tradisional (simplisia untuk skala industri dan peningkatan produksi tanaman obat dalam skala besar menjadi tidak ekonomis tanpa pestisida. Disatu sisi penggunaan pestisida dapat menguntungkan yaitu menyebabkan toksis pada hama namun disisi lain toksisitas dapat terjadi juga pada manusia, sehingga residu pestisida dalam tanaman obat yang dikonsumsi dalam jangka panjang akan merugikan kesehatan. Batas maksimum Residu (BMR pestisida dalam simplisia baik di Indonesia maupun di negara lain belum ditetapkan. Sehingga untuk itu untuk mengetahui adanya residu pestisida jenis organoklorin yang telah dilarang penggunaannya melalui Permentan No.434.1/kpts/TP.270/7/2001 dan untuk mengetahui batas keamanannya, maka perlu dilakukan penetapan residu organoklorin dalam simplisia dan menetapkan batas keamanan berdasarkan perhitungan secara teoritis. Pengujian residu dilakukan terhadap golongan pestisida organoklorin pada 4 jenis simplisia (daun wungu (Graptophyllum pictum (L Grifl, daun sambiloto Andrographis paniculata Ness, herba pegagan (Centella  asiatica (L Urban, daun tempuyung (Sonchus arvensis (L yang berasal dari 3 lokasi penanaman, yaitu : perkebunan Tanaman Obat Manako (Jawa Barat, Balai Penelitian Tanaman Obat Tawangmangu (BPTO di Jawa Tengah dan Perkebunan Tanaman Obat Purwodadi (Jawa Timur. Pemeriksaan residu pestisida organoklorin menggunakan kromatografi gas dan perhitungan batas keamanan dihitung dengan adanya nilai ADI (Acceptable daily intake yang telah ditetapkan bersama antara JAO dan WHO serta perkiraan banyaknya konsumsi simplisia. Hasil Pengujian residu pestisida organoklorin diperoleh bahwa simplisia daun Wungu (Tawangmangu mengandung residu lindan dengan kadar 0,24 mg/kg, pegagan (Purwodadi, mengandung lindan 0,36 mg/kg dan aldrin 0,31 mg/kg serta pegagan (Manako mengandung heptaklor 0,15 mg/kg dan op-DDE 0,11 mg/kg. Adapun penghitungan BMR heptaklor dan lindan secara teoritis dengan asumsi rata

  8. Algorithmic mathematics

    CERN Document Server

    Hougardy, Stefan

    2016-01-01

    Algorithms play an increasingly important role in nearly all fields of mathematics. This book allows readers to develop basic mathematical abilities, in particular those concerning the design and analysis of algorithms as well as their implementation. It presents not only fundamental algorithms like the sieve of Eratosthenes, the Euclidean algorithm, sorting algorithms, algorithms on graphs, and Gaussian elimination, but also discusses elementary data structures, basic graph theory, and numerical questions. In addition, it provides an introduction to programming and demonstrates in detail how to implement algorithms in C++. This textbook is suitable for students who are new to the subject and covers a basic mathematical lecture course, complementing traditional courses on analysis and linear algebra. Both authors have given this "Algorithmic Mathematics" course at the University of Bonn several times in recent years.

  9. Residual magnetic field in rotary machines; Campo magnetico residual en maquinas rotatorias

    Energy Technology Data Exchange (ETDEWEB)

    Gonzalez V, Esteban A; Apanco R, Marcelino [Instituto de Investigaciones Electricas, Cuernavaca, Morelos (Mexico)

    2007-07-01

    The residual magnetism is a phenomenon in which the magnetic dipoles of a substance are oriented in a certain degree. On the other hand, when internal forces exist capable of aligning elementary magnetic dipoles of a material, a permanent magnet is obtained. Just as in a conductor or in a material, in the elements of a rotary electrical machine magnetic fields can be induced that produce a residual magnetism or magnetization. In the rotary electrical machines, the magnetization phenomenon causes serious problems, such as the generation of induced currents that propitiate the mechanical wear in bearings, collars, trunnions and inclusive in the shaft, by effects known as pitting, frosting and spark tracks, as well as erroneous readings in vibration and temperature sensors, that in some cases can cause the shut down of the machine. In this article are presented the general concepts on the residual magnetism in rotary electrical machines, the causes that originate it and the problems that arises, as well as the demagnetization of the components that have residual magnetic field. The results obtained by the area of Electrical Equipment of the Instituto de Investigaciones Electricas are revised, during the execution of activities related to the measurement and elimination of the residual magnetic field in rotary electrical machines. [Spanish] El magnetismo residual es un fenomeno en el que los dipolos magneticos de una sustancia se encuentran orientados en un grado determinado. Por otro lado, cuando existen fuerzas internas capaces de alinear los dipolos magneticos elementales de un material, se tiene un iman permanente. Al igual que en un conductor o un material, en los elementos de una maquina electrica rotatoria se pueden inducir campos magneticos que producen un magnetismo residual o magnetizacion. En las maquinas electricas rotatorias, el fenomeno de magnetizacion causa graves problemas, como la generacion de corrientes inducidas que propician el desgaste mecanico

  10. Optimisation of process parameters in friction stir welding based on residual stress analysis: a feasibility study

    DEFF Research Database (Denmark)

    Tutum, Cem Celal; Hattel, Jesper Henri

    2010-01-01

    The present paper considers the optimisation of process parameters in friction stir welding (FSW). More specifically, the choices of rotational speed and traverse welding speed have been investigated using genetic algorithms. The welding process is simulated in a transient, two......-dimensional sequentially coupled thermomechanical model in ANSYS. This model is then used in an optimisation case where the two objectives are the minimisation of the peak residual stresses and the maximisation of the welding speed. The results indicate that the objectives for the considered case are conflicting......, and this is presented as a Pareto optimal front. Moreover, a higher welding speed for a fixed rotational speed results, in general, in slightly higher stress levels in the tension zone, whereas a higher rotational speed for a fixed welding speed yields somewhat lower peak residual stress, however, a wider tension zone...

  11. Essential algorithms a practical approach to computer algorithms

    CERN Document Server

    Stephens, Rod

    2013-01-01

    A friendly and accessible introduction to the most useful algorithms Computer algorithms are the basic recipes for programming. Professional programmers need to know how to use algorithms to solve difficult programming problems. Written in simple, intuitive English, this book describes how and when to use the most practical classic algorithms, and even how to create new algorithms to meet future needs. The book also includes a collection of questions that can help readers prepare for a programming job interview. Reveals methods for manipulating common data structures s

  12. Algorithmic cryptanalysis

    CERN Document Server

    Joux, Antoine

    2009-01-01

    Illustrating the power of algorithms, Algorithmic Cryptanalysis describes algorithmic methods with cryptographically relevant examples. Focusing on both private- and public-key cryptographic algorithms, it presents each algorithm either as a textual description, in pseudo-code, or in a C code program.Divided into three parts, the book begins with a short introduction to cryptography and a background chapter on elementary number theory and algebra. It then moves on to algorithms, with each chapter in this section dedicated to a single topic and often illustrated with simple cryptographic applic

  13. Extending the eigCG algorithm to nonsymmetric Lanczos for linear systems with multiple right-hand sides

    Energy Technology Data Exchange (ETDEWEB)

    Abdel-Rehim, A M; Stathopoulos, Andreas; Orginos, Kostas

    2014-08-01

    The technique that was used to build the EigCG algorithm for sparse symmetric linear systems is extended to the nonsymmetric case using the BiCG algorithm. We show that, similarly to the symmetric case, we can build an algorithm that is capable of computing a few smallest magnitude eigenvalues and their corresponding left and right eigenvectors of a nonsymmetric matrix using only a small window of the BiCG residuals while simultaneously solving a linear system with that matrix. For a system with multiple right-hand sides, we give an algorithm that computes incrementally more eigenvalues while solving the first few systems and then uses the computed eigenvectors to deflate BiCGStab for the remaining systems. Our experiments on various test problems, including Lattice QCD, show the remarkable ability of EigBiCG to compute spectral approximations with accuracy comparable to that of the unrestarted, nonsymmetric Lanczos. Furthermore, our incremental EigBiCG followed by appropriately restarted and deflated BiCGStab provides a competitive method for systems with multiple right-hand sides.

  14. Mapping allostery through computational glycine scanning and correlation analysis of residue-residue contacts.

    Science.gov (United States)

    Johnson, Quentin R; Lindsay, Richard J; Nellas, Ricky B; Fernandez, Elias J; Shen, Tongye

    2015-02-24

    Understanding allosteric mechanisms is essential for the physical control of molecular switches and downstream cellular responses. However, it is difficult to decode essential allosteric motions in a high-throughput scheme. A general two-pronged approach to performing automatic data reduction of simulation trajectories is presented here. The first step involves coarse-graining and identifying the most dynamic residue-residue contacts. The second step is performing principal component analysis of these contacts and extracting the large-scale collective motions expressed via these residue-residue contacts. We demonstrated the method using a protein complex of nuclear receptors. Using atomistic modeling and simulation, we examined the protein complex and a set of 18 glycine point mutations of residues that constitute the binding pocket of the ligand effector. The important motions that are responsible for the allostery are reported. In contrast to conventional induced-fit and lock-and-key binding mechanisms, a novel "frustrated-fit" binding mechanism of RXR for allosteric control was revealed.

  15. Characterization Report on Sand, Slag, and Crucible Residues and on Fluoride Residues

    International Nuclear Information System (INIS)

    Murray, A.M.

    1999-01-01

    This paper reports on the chemical characterization of the sand, slag, and crucible (SS and C) residues and the fluoride residues that may be shipped from the Rocky Flats Environmental Technology Site (RFETS) to Savannah River Site (SRS)

  16. Denni Algorithm An Enhanced Of SMS (Scan, Move and Sort) Algorithm

    Science.gov (United States)

    Aprilsyah Lubis, Denni; Salim Sitompul, Opim; Marwan; Tulus; Andri Budiman, M.

    2017-12-01

    Sorting has been a profound area for the algorithmic researchers, and many resources are invested to suggest a more working sorting algorithm. For this purpose many existing sorting algorithms were observed in terms of the efficiency of the algorithmic complexity. Efficient sorting is important to optimize the use of other algorithms that require sorted lists to work correctly. Sorting has been considered as a fundamental problem in the study of algorithms that due to many reasons namely, the necessary to sort information is inherent in many applications, algorithms often use sorting as a key subroutine, in algorithm design there are many essential techniques represented in the body of sorting algorithms, and many engineering issues come to the fore when implementing sorting algorithms., Many algorithms are very well known for sorting the unordered lists, and one of the well-known algorithms that make the process of sorting to be more economical and efficient is SMS (Scan, Move and Sort) algorithm, an enhancement of Quicksort invented Rami Mansi in 2010. This paper presents a new sorting algorithm called Denni-algorithm. The Denni algorithm is considered as an enhancement on the SMS algorithm in average, and worst cases. The Denni algorithm is compared with the SMS algorithm and the results were promising.

  17. Identification of Arbitrary Zonation in Groundwater Parameters using the Level Set Method and a Parallel Genetic Algorithm

    Science.gov (United States)

    Lei, H.; Lu, Z.; Vesselinov, V. V.; Ye, M.

    2017-12-01

    Simultaneous identification of both the zonation structure of aquifer heterogeneity and the hydrogeological parameters associated with these zones is challenging, especially for complex subsurface heterogeneity fields. In this study, a new approach, based on the combination of the level set method and a parallel genetic algorithm is proposed. Starting with an initial guess for the zonation field (including both zonation structure and the hydraulic properties of each zone), the level set method ensures that material interfaces are evolved through the inverse process such that the total residual between the simulated and observed state variables (hydraulic head) always decreases, which means that the inversion result depends on the initial guess field and the minimization process might fail if it encounters a local minimum. To find the global minimum, the genetic algorithm (GA) is utilized to explore the parameters that define initial guess fields, and the minimal total residual corresponding to each initial guess field is considered as the fitness function value in the GA. Due to the expensive evaluation of the fitness function, a parallel GA is adapted in combination with a simulated annealing algorithm. The new approach has been applied to several synthetic cases in both steady-state and transient flow fields, including a case with real flow conditions at the chromium contaminant site at the Los Alamos National Laboratory. The results show that this approach is capable of identifying the arbitrary zonation structures of aquifer heterogeneity and the hydrogeological parameters associated with these zones effectively.

  18. Improved k-t PCA Algorithm Using Artificial Sparsity in Dynamic MRI.

    Science.gov (United States)

    Wang, Yiran; Chen, Zhifeng; Wang, Jing; Yuan, Lixia; Xia, Ling; Liu, Feng

    2017-01-01

    The k - t principal component analysis ( k - t PCA) is an effective approach for high spatiotemporal resolution dynamic magnetic resonance (MR) imaging. However, it suffers from larger residual aliasing artifacts and noise amplification when the reduction factor goes higher. To further enhance the performance of this technique, we propose a new method called sparse k - t PCA that combines the k - t PCA algorithm with an artificial sparsity constraint. It is a self-calibrated procedure that is based on the traditional k - t PCA method by further eliminating the reconstruction error derived from complex subtraction of the sampled k - t space from the original reconstructed k - t space. The proposed method is tested through both simulations and in vivo datasets with different reduction factors. Compared to the standard k - t PCA algorithm, the sparse k - t PCA can improve the normalized root-mean-square error performance and the accuracy of temporal resolution. It is thus useful for rapid dynamic MR imaging.

  19. Python algorithms mastering basic algorithms in the Python language

    CERN Document Server

    Hetland, Magnus Lie

    2014-01-01

    Python Algorithms, Second Edition explains the Python approach to algorithm analysis and design. Written by Magnus Lie Hetland, author of Beginning Python, this book is sharply focused on classical algorithms, but it also gives a solid understanding of fundamental algorithmic problem-solving techniques. The book deals with some of the most important and challenging areas of programming and computer science in a highly readable manner. It covers both algorithmic theory and programming practice, demonstrating how theory is reflected in real Python programs. Well-known algorithms and data struc

  20. Residual stress by repair welds

    International Nuclear Information System (INIS)

    Mochizuki, Masahito; Toyoda, Masao

    2003-01-01

    Residual stress by repair welds is computed using the thermal elastic-plastic analysis with phase-transformation effect. Coupling phenomena of temperature, microstructure, and stress-strain fields are simulated in the finite-element analysis. Weld bond of a plate butt-welded joint is gouged and then deposited by weld metal in repair process. Heat source is synchronously moved with the deposition of the finite-element as the weld deposition. Microstructure is considered by using CCT diagram and the transformation behavior in the repair weld is also simulated. The effects of initial stress, heat input, and weld length on residual stress distribution are studied from the organic results of numerical analysis. Initial residual stress before repair weld has no influence on the residual stress after repair treatment near weld metal, because the initial stress near weld metal releases due to high temperature of repair weld and then stress by repair weld regenerates. Heat input has an effect for residual stress distribution, for not its magnitude but distribution zone. Weld length should be considered reducing the magnitude of residual stress in the edge of weld bead; short bead induces high tensile residual stress. (author)

  1. An Energy Efficient Stable Election-Based Routing Algorithm for Wireless Sensor Networks

    Directory of Open Access Journals (Sweden)

    Weiwei Yuan

    2013-10-01

    Full Text Available Sensor nodes usually have limited energy supply and they are impractical to recharge. How to balance traffic load in sensors in order to increase network lifetime is a very challenging research issue. Many clustering algorithms have been proposed recently for wireless sensor networks (WSNs. However, sensor networks with one fixed sink node often suffer from a hot spots problem since nodes near sinks have more traffic burden to forward during a multi-hop transmission process. The use of mobile sinks has been shown to be an effective technique to enhance network performance features such as latency, energy efficiency, network lifetime, etc. In this paper, a modified Stable Election Protocol (SEP, which employs a mobile sink, has been proposed for WSNs with non-uniform node distribution. The decision of selecting cluster heads by the sink is based on the minimization of the associated additional energy and residual energy at each node. Besides, the cluster head selects the shortest path to reach the sink between the direct approach and the indirect approach with the use of the nearest cluster head. Simulation results demonstrate that our algorithm has better performance than traditional routing algorithms, such as LEACH and SEP.

  2. Reactivity of Athabasca residue and of its SARA fractions during residue hydroconversion

    Energy Technology Data Exchange (ETDEWEB)

    Verstraete, J.; Danial-Fortain, P.; Gauthier, T.; Merdrignac, I. [IFP-Lyon, Vermaison (France); Budzinski, H. [Bordeaux Univ. (France). ISM-LPTC, UMR CNRS

    2009-07-01

    Residue conversion processes are becoming increasingly important because of the declining market for residual fuel oil and a greater demand for middle distillates. Ebullated-bed hydroconversion is a commercially proven technology for converting heavy feedstocks with high amounts of impurities. The process enables the conversion of atmospheric or vacuum residues at temperatures up to 440 degrees C, and at liquid hourly space velocity (LHSV) conditions in the range of 0.15 to 0.5 per hour. A 540 degrees C conversion of up to 80 weight per cent can be achieved under these conditions. This paper reported on a research study conducted at IFP Lyon in which the residue hydroconversion in a large-scale ebullated bed bench unit was investigated to determine the impact of operating conditions and feed properties on yield and product qualities. Hydrogen was added to the feed in the bench units to keep a high hydrogen partial pressure and favour the catalytic hydroconversion reactions. In a typical test, the reactor was fed with 50 g of feedstock and 0.45 g of crushed equilibrium industrial NiMo catalyst, pressurized hydrogen and quickly heated at the reaction temperature. This paper also discussed the conversion of Athabasca bitumen residue in the large-scale pilot plant and also in the small scale batch reactor. The effect of operating temperature and space velocity was examined. The reactivity of the saturates, aromatics, resins and asphaltenes (SARA) fractions of the bitumen was studied separately in order to better understand the conversion mechanisms and reactivities. The Athabasca bitumen feed and SARA fractions were also analyzed in terms of standard petroleum analysis, SARA fractionation, elemental analysis, size exclusion chromatography (SEC) and 13C NMR. Hydroconversion experiments were conducted in the batch unit at different reaction temperatures and reaction times. A comparison of small-scale batch results with those obtained with the continuous large-scale bench

  3. An analytical study of composite laminate lay-up using search algorithms for maximization of flexural stiffness and minimization of springback angle

    Science.gov (United States)

    Singh, Ranjan Kumar; Rinawa, Moti Lal

    2018-04-01

    The residual stresses arising in fiber-reinforced laminates during their curing in closed molds lead to changes in the composites after their removal from the molds and cooling. One of these dimensional changes of angle sections is called springback. The parameters such as lay-up, stacking sequence, material system, cure temperature, thickness etc play important role in it. In present work, it is attempted to optimize lay-up and stacking sequence for maximization of flexural stiffness and minimization of springback angle. The search algorithms are employed to obtain best sequence through repair strategy such as swap. A new search algorithm, termed as lay-up search algorithm (LSA) is also proposed, which is an extension of permutation search algorithm (PSA). The efficacy of PSA and LSA is tested on the laminates with a range of lay-ups. A computer code is developed on MATLAB implementing the above schemes. Also, the strategies for multi objective optimization using search algorithms are suggested and tested.

  4. Hybrid Cryptosystem Using Tiny Encryption Algorithm and LUC Algorithm

    Science.gov (United States)

    Rachmawati, Dian; Sharif, Amer; Jaysilen; Andri Budiman, Mohammad

    2018-01-01

    Security becomes a very important issue in data transmission and there are so many methods to make files more secure. One of that method is cryptography. Cryptography is a method to secure file by writing the hidden code to cover the original file. Therefore, if the people do not involve in cryptography, they cannot decrypt the hidden code to read the original file. There are many methods are used in cryptography, one of that method is hybrid cryptosystem. A hybrid cryptosystem is a method that uses a symmetric algorithm to secure the file and use an asymmetric algorithm to secure the symmetric algorithm key. In this research, TEA algorithm is used as symmetric algorithm and LUC algorithm is used as an asymmetric algorithm. The system is tested by encrypting and decrypting the file by using TEA algorithm and using LUC algorithm to encrypt and decrypt the TEA key. The result of this research is by using TEA Algorithm to encrypt the file, the cipher text form is the character from ASCII (American Standard for Information Interchange) table in the form of hexadecimal numbers and the cipher text size increase by sixteen bytes as the plaintext length is increased by eight characters.

  5. Simultaneous determination of aquifer parameters and zone structures with fuzzy c-means clustering and meta-heuristic harmony search algorithm

    Science.gov (United States)

    Ayvaz, M. Tamer

    2007-11-01

    This study proposes an inverse solution algorithm through which both the aquifer parameters and the zone structure of these parameters can be determined based on a given set of observations on piezometric heads. In the zone structure identification problem fuzzy c-means ( FCM) clustering method is used. The association of the zone structure with the transmissivity distribution is accomplished through an optimization model. The meta-heuristic harmony search ( HS) algorithm, which is conceptualized using the musical process of searching for a perfect state of harmony, is used as an optimization technique. The optimum parameter zone structure is identified based on three criteria which are the residual error, parameter uncertainty, and structure discrimination. A numerical example given in the literature is solved to demonstrate the performance of the proposed algorithm. Also, a sensitivity analysis is performed to test the performance of the HS algorithm for different sets of solution parameters. Results indicate that the proposed solution algorithm is an effective way in the simultaneous identification of aquifer parameters and their corresponding zone structures.

  6. Optimization of the Process Parameters for Controlling Residual Stress and Distortion in Friction Stir Welding

    DEFF Research Database (Denmark)

    Tutum, Cem Celal; Schmidt, Henrik Nikolaj Blicher; Hattel, Jesper Henri

    2008-01-01

    In the present paper, numerical optimization of the process parameters, i.e. tool rotation speed and traverse speed, aiming minimization of the two conflicting objectives, i.e. the residual stresses and welding time, subjected to process-specific thermal constraints in friction stir welding......, is investigated. The welding process is simulated in 2-dimensions with a sequentially coupled transient thermo-mechanical model using ANSYS. The numerical optimization problem is implemented in modeFRONTIER and solved using the Multi-Objective Genetic Algorithm (MOGA-II). An engineering-wise evaluation or ranking...

  7. Sound algorithms

    OpenAIRE

    De Götzen , Amalia; Mion , Luca; Tache , Olivier

    2007-01-01

    International audience; We call sound algorithms the categories of algorithms that deal with digital sound signal. Sound algorithms appeared in the very infancy of computer. Sound algorithms present strong specificities that are the consequence of two dual considerations: the properties of the digital sound signal itself and its uses, and the properties of auditory perception.

  8. An improved algorithm for the determination of the system paramters of a visual binary by least squares

    Science.gov (United States)

    Xu, Yu-Lin

    The problem of computing the orbit of a visual binary from a set of observed positions is reconsidered. It is a least squares adjustment problem, if the observational errors follow a bias-free multivariate Gaussian distribution and the covariance matrix of the observations is assumed to be known. The condition equations are constructed to satisfy both the conic section equation and the area theorem, which are nonlinear in both the observations and the adjustment parameters. The traditional least squares algorithm, which employs condition equations that are solved with respect to the uncorrelated observations and either linear in the adjustment parameters or linearized by developing them in Taylor series by first-order approximation, is inadequate in our orbit problem. D.C. Brown proposed an algorithm solving a more general least squares adjustment problem in which the scalar residual function, however, is still constructed by first-order approximation. Not long ago, a completely general solution was published by W.H Jefferys, who proposed a rigorous adjustment algorithm for models in which the observations appear nonlinearly in the condition equations and may be correlated, and in which construction of the normal equations and the residual function involves no approximation. This method was successfully applied in our problem. The normal equations were first solved by Newton's scheme. Practical examples show that this converges fast if the observational errors are sufficiently small and the initial approximate solution is sufficiently accurate, and that it fails otherwise. Newton's method was modified to yield a definitive solution in the case the normal approach fails, by combination with the method of steepest descent and other sophisticated algorithms. Practical examples show that the modified Newton scheme can always lead to a final solution. The weighting of observations, the orthogonal parameters and the efficiency of a set of adjustment parameters are also

  9. SU-D-17A-02: Four-Dimensional CBCT Using Conventional CBCT Dataset and Iterative Subtraction Algorithm of a Lung Patient

    International Nuclear Information System (INIS)

    Hu, E; Lasio, G; Yi, B

    2014-01-01

    Purpose: The Iterative Subtraction Algorithm (ISA) method generates retrospectively a pre-selected motion phase cone-beam CT image from the full motion cone-beam CT acquired at standard rotation speed. This work evaluates ISA method with real lung patient data. Methods: The goal of the ISA algorithm is to extract motion and no- motion components form the full reconstruction CBCT. The workflow consists of subtracting from the full CBCT all of the undesired motion phases and obtain a motion de-blurred single-phase CBCT image, followed by iteration of this subtraction process. ISA is realized as follows: 1) The projections are sorted to various phases, and from all phases, a full reconstruction is performed to generate an image CTM. 2) Generate forward projections of CTM at the desired phase projection angles, the subtraction of projection and the forward projection will reconstruct a CTSub1, which diminishes the desired phase component. 3) By adding back the CTSub1 to CTm, no motion CBCT, CTS1, can be computed. 4) CTS1 still contains residual motion component. 5) This residual motion component can be further reduced by iteration.The ISA 4DCBCT technique was implemented using Varian Trilogy accelerator OBI system. To evaluate the method, a lung patient CBCT dataset was used. The reconstruction algorithm is FDK. Results: The single phase CBCT reconstruction generated via ISA successfully isolates the desired motion phase from the full motion CBCT, effectively reducing motion blur. It also shows improved image quality, with reduced streak artifacts with respect to the reconstructions from unprocessed phase-sorted projections only. Conclusion: A CBCT motion de-blurring algorithm, ISA, has been developed and evaluated with lung patient data. The algorithm allows improved visualization of a single phase motion extracted from a standard CBCT dataset. This study has been supported by National Institute of Health through R01CA133539

  10. Defining an essence of structure determining residue contacts in proteins.

    Science.gov (United States)

    Sathyapriya, R; Duarte, Jose M; Stehr, Henning; Filippis, Ioannis; Lappe, Michael

    2009-12-01

    The network of native non-covalent residue contacts determines the three-dimensional structure of a protein. However, not all contacts are of equal structural significance, and little knowledge exists about a minimal, yet sufficient, subset required to define the global features of a protein. Characterisation of this "structural essence" has remained elusive so far: no algorithmic strategy has been devised to-date that could outperform a random selection in terms of 3D reconstruction accuracy (measured as the Ca RMSD). It is not only of theoretical interest (i.e., for design of advanced statistical potentials) to identify the number and nature of essential native contacts-such a subset of spatial constraints is very useful in a number of novel experimental methods (like EPR) which rely heavily on constraint-based protein modelling. To derive accurate three-dimensional models from distance constraints, we implemented a reconstruction pipeline using distance geometry. We selected a test-set of 12 protein structures from the four major SCOP fold classes and performed our reconstruction analysis. As a reference set, series of random subsets (ranging from 10% to 90% of native contacts) are generated for each protein, and the reconstruction accuracy is computed for each subset. We have developed a rational strategy, termed "cone-peeling" that combines sequence features and network descriptors to select minimal subsets that outperform the reference sets. We present, for the first time, a rational strategy to derive a structural essence of residue contacts and provide an estimate of the size of this minimal subset. Our algorithm computes sparse subsets capable of determining the tertiary structure at approximately 4.8 A Ca RMSD with as little as 8% of the native contacts (Ca-Ca and Cb-Cb). At the same time, a randomly chosen subset of native contacts needs about twice as many contacts to reach the same level of accuracy. This "structural essence" opens new avenues in the

  11. Optimization of IBF parameters based on adaptive tool-path algorithm

    Science.gov (United States)

    Deng, Wen Hui; Chen, Xian Hua; Jin, Hui Liang; Zhong, Bo; Hou, Jin; Li, An Qi

    2018-03-01

    As a kind of Computer Controlled Optical Surfacing(CCOS) technology. Ion Beam Figuring(IBF) has obvious advantages in the control of surface accuracy, surface roughness and subsurface damage. The superiority and characteristics of IBF in optical component processing are analyzed from the point of view of removal mechanism. For getting more effective and automatic tool path with the information of dwell time, a novel algorithm is proposed in this thesis. Based on the removal functions made through our IBF equipment and the adaptive tool-path, optimized parameters are obtained through analysis the residual error that would be created in the polishing process. A Φ600 mm plane reflector element was used to be a simulation instance. The simulation result shows that after four combinations of processing, the surface accuracy of PV (Peak Valley) value and the RMS (Root Mean Square) value was reduced to 4.81 nm and 0.495 nm from 110.22 nm and 13.998 nm respectively in the 98% aperture. The result shows that the algorithm and optimized parameters provide a good theoretical for high precision processing of IBF.

  12. Residuals and the Residual-Based Statistic for Testing Goodness of Fit of Structural Equation Models

    Science.gov (United States)

    Foldnes, Njal; Foss, Tron; Olsson, Ulf Henning

    2012-01-01

    The residuals obtained from fitting a structural equation model are crucial ingredients in obtaining chi-square goodness-of-fit statistics for the model. The authors present a didactic discussion of the residuals, obtaining a geometrical interpretation by recognizing the residuals as the result of oblique projections. This sheds light on the…

  13. Genetic algorithms

    Science.gov (United States)

    Wang, Lui; Bayer, Steven E.

    1991-01-01

    Genetic algorithms are mathematical, highly parallel, adaptive search procedures (i.e., problem solving methods) based loosely on the processes of natural genetics and Darwinian survival of the fittest. Basic genetic algorithms concepts are introduced, genetic algorithm applications are introduced, and results are presented from a project to develop a software tool that will enable the widespread use of genetic algorithm technology.

  14. Machine for compacting solid residues

    International Nuclear Information System (INIS)

    Herzog, J.

    1981-11-01

    Machine for compacting solid residues, particularly bulky radioactive residues, constituted of a horizontally actuated punch and a fixed compression anvil, in which the residues are first compacted horizontally and then vertically. Its salient characteristic is that the punch and the compression anvil have embossments on the compression side and interpenetrating plates in the compression position [fr

  15. Multi-level learning: improving the prediction of protein, domain and residue interactions by allowing information flow between levels

    Directory of Open Access Journals (Sweden)

    McDermott Drew

    2009-08-01

    Full Text Available Abstract Background Proteins interact through specific binding interfaces that contain many residues in domains. Protein interactions thus occur on three different levels of a concept hierarchy: whole-proteins, domains, and residues. Each level offers a distinct and complementary set of features for computationally predicting interactions, including functional genomic features of whole proteins, evolutionary features of domain families and physical-chemical features of individual residues. The predictions at each level could benefit from using the features at all three levels. However, it is not trivial as the features are provided at different granularity. Results To link up the predictions at the three levels, we propose a multi-level machine-learning framework that allows for explicit information flow between the levels. We demonstrate, using representative yeast interaction networks, that our algorithm is able to utilize complementary feature sets to make more accurate predictions at the three levels than when the three problems are approached independently. To facilitate application of our multi-level learning framework, we discuss three key aspects of multi-level learning and the corresponding design choices that we have made in the implementation of a concrete learning algorithm. 1 Architecture of information flow: we show the greater flexibility of bidirectional flow over independent levels and unidirectional flow; 2 Coupling mechanism of the different levels: We show how this can be accomplished via augmenting the training sets at each level, and discuss the prevention of error propagation between different levels by means of soft coupling; 3 Sparseness of data: We show that the multi-level framework compounds data sparsity issues, and discuss how this can be dealt with by building local models in information-rich parts of the data. Our proof-of-concept learning algorithm demonstrates the advantage of combining levels, and opens up

  16. A generic library for large scale solution of PDEs on modern heterogeneous architectures

    DEFF Research Database (Denmark)

    Glimberg, Stefan Lemvig; Engsig-Karup, Allan Peter

    2012-01-01

    Adapting to new programming models for modern multi- and many-core architectures requires code-rewriting and changing algorithms and data structures, in order to achieve good efficiency and scalability. We present a generic library for solving large scale partial differential equations (PDEs......), capable of utilizing heterogeneous CPU/GPU environments. The library can be used for fast proto-typing of PDE solvers, based on finite difference approximations of spatial derivatives in one, two, or three dimensions. In order to efficiently solve large scale problems, we keep memory consumption...... and memory access low, using a low-storage implementation of flexible-order finite difference operators. We will illustrate the use of library components by assembling such matrix-free operators to be used with one of the supported iterative solvers, such as GMRES, CG, Multigrid or Defect Correction...

  17. An evaluation of parallel multigrid as a solver and a preconditioner for singular perturbed problems

    Energy Technology Data Exchange (ETDEWEB)

    Oosterlee, C.W. [Inst. for Algorithms and Scientific Computing, Sankt Augustin (Germany); Washio, T. [C& C Research Lab., Sankt Augustin (Germany)

    1996-12-31

    In this paper we try to achieve h-independent convergence with preconditioned GMRES and BiCGSTAB for 2D singular perturbed equations. Three recently developed multigrid methods are adopted as a preconditioner. They are also used as solution methods in order to compare the performance of the methods as solvers and as preconditioners. Two of the multigrid methods differ only in the transfer operators. One uses standard matrix- dependent prolongation operators from. The second uses {open_quotes}upwind{close_quotes} prolongation operators, developed. Both employ the Galerkin coarse grid approximation and an alternating zebra line Gauss-Seidel smoother. The third method is based on the block LU decomposition of a matrix and on an approximate Schur complement. This multigrid variant is presented in. All three multigrid algorithms are algebraic methods.

  18. Algorithm aversion: people erroneously avoid algorithms after seeing them err.

    Science.gov (United States)

    Dietvorst, Berkeley J; Simmons, Joseph P; Massey, Cade

    2015-02-01

    Research shows that evidence-based algorithms more accurately predict the future than do human forecasters. Yet when forecasters are deciding whether to use a human forecaster or a statistical algorithm, they often choose the human forecaster. This phenomenon, which we call algorithm aversion, is costly, and it is important to understand its causes. We show that people are especially averse to algorithmic forecasters after seeing them perform, even when they see them outperform a human forecaster. This is because people more quickly lose confidence in algorithmic than human forecasters after seeing them make the same mistake. In 5 studies, participants either saw an algorithm make forecasts, a human make forecasts, both, or neither. They then decided whether to tie their incentives to the future predictions of the algorithm or the human. Participants who saw the algorithm perform were less confident in it, and less likely to choose it over an inferior human forecaster. This was true even among those who saw the algorithm outperform the human.

  19. The Texas Medication Algorithm Project (TMAP) schizophrenia algorithms.

    Science.gov (United States)

    Miller, A L; Chiles, J A; Chiles, J K; Crismon, M L; Rush, A J; Shon, S P

    1999-10-01

    In the Texas Medication Algorithm Project (TMAP), detailed guidelines for medication management of schizophrenia and related disorders, bipolar disorders, and major depressive disorders have been developed and implemented. This article describes the algorithms developed for medication treatment of schizophrenia and related disorders. The guidelines recommend a sequence of medications and discuss dosing, duration, and switch-over tactics. They also specify response criteria at each stage of the algorithm for both positive and negative symptoms. The rationale and evidence for each aspect of the algorithms are presented.

  20. Residual stress concerns in containment analysis

    International Nuclear Information System (INIS)

    Costantini, F.; Kulak, R. F.; Pfeiffer, P. A.

    1997-01-01

    The manufacturing of steel containment vessels starts with the forming of flat plates into curved plates. A steel containment structure is made by welding individual plates together to form the sections that make up the complex shaped vessels. The metal forming and welding process leaves residual stresses in the vessel walls. Generally, the effect of metal forming residual stresses can be reduced or virtually eliminated by thermally stress relieving the vesseL In large containment vessels this may not be practical and thus the residual stresses due to manufacturing may become important. The residual stresses could possibly tiect the response of the vessel to internal pressurization. When the level of residual stresses is significant it will affect the vessel's response, for instance the yielding pressure and possibly the failure pressure. The paper will address the effect of metal forming residual stresses on the response of a generic pressure vessel to internal pressurization. A scoping analysis investigated the effect of residual forming stresses on the response of an internally pressurized vessel. A simple model was developed to gain understanding of the mechanics of the problem. Residual stresses due to the welding process were not considered in this investigation

  1. Dose calculations algorithm for narrow heavy charged-particle beams

    Energy Technology Data Exchange (ETDEWEB)

    Barna, E A; Kappas, C [Department of Medical Physics, School of Medicine, University of Patras (Greece); Scarlat, F [National Institute for Laser and Plasma Physics, Bucharest (Romania)

    1999-12-31

    The dose distributional advantages of the heavy charged-particles can be fully exploited by using very efficient and accurate dose calculation algorithms, which can generate optimal three-dimensional scanning patterns. An inverse therapy planning algorithm for dynamically scanned, narrow heavy charged-particle beams is presented in this paper. The irradiation `start point` is defined at the distal end of the target volume, right-down, in a beam`s eye view. The peak-dose of the first elementary beam is set to be equal to the prescribed dose in the target volume, and is defined as the reference dose. The weighting factor of any Bragg-peak is determined by the residual dose at the point of irradiation, calculated as the difference between the reference dose and the cumulative dose delivered at that point of irradiation by all the previous Bragg-peaks. The final pattern consists of the weighted Bragg-peaks irradiation density. Dose distributions were computed using two different scanning steps equal to 0.5 mm, and 1 mm respectively. Very accurate and precise localized dose distributions, conform to the target volume, were obtained. (authors) 6 refs., 3 figs.

  2. Characterization of Hospital Residuals

    International Nuclear Information System (INIS)

    Blanco Meza, A.; Bonilla Jimenez, S.

    1997-01-01

    The main objective of this investigation is the characterization of the solid residuals. A description of the handling of the liquid and gassy waste generated in hospitals is also given, identifying the source where they originate. To achieve the proposed objective the work was divided in three stages: The first one was the planning and the coordination with each hospital center, in this way, to determine the schedule of gathering of the waste can be possible. In the second stage a fieldwork was made; it consisted in gathering the quantitative and qualitative information of the general state of the handling of residuals. In the third and last stage, the information previously obtained was organized to express the results as the production rate per day by bed, generation of solid residuals for sampled services, type of solid residuals and density of the same ones. With the obtained results, approaches are settled down to either determine design parameters for final disposition whether for incineration, trituration, sanitary filler or recycling of some materials, and storage politics of the solid residuals that allow to determine the gathering frequency. The study concludes that it is necessary to improve the conditions of the residuals handling in some aspects, to provide the cleaning personnel of the equipment for gathering disposition and of security, minimum to carry out this work efficiently, and to maintain a control of all the dangerous waste, like sharp or polluted materials. In this way, an appreciable reduction is guaranteed in the impact on the atmosphere. (Author) [es

  3. A new algorithm for construction of coarse-grained sites of large biomolecules.

    Science.gov (United States)

    Li, Min; Zhang, John Z H; Xia, Fei

    2016-04-05

    The development of coarse-grained (CG) models for large biomolecules remains a challenge in multiscale simulations, including a rigorous definition of CG representations for them. In this work, we proposed a new stepwise optimization imposed with the boundary-constraint (SOBC) algorithm to construct the CG sites of large biomolecules, based on the s cheme of essential dynamics CG. By means of SOBC, we can rigorously derive the CG representations of biomolecules with less computational cost. The SOBC is particularly efficient for the CG definition of large systems with thousands of residues. The resulted CG sites can be parameterized as a CG model using the normal mode analysis based fluctuation matching method. Through normal mode analysis, the obtained modes of CG model can accurately reflect the functionally related slow motions of biomolecules. The SOBC algorithm can be used for the construction of CG sites of large biomolecules such as F-actin and for the study of mechanical properties of biomaterials. © 2015 Wiley Periodicals, Inc.

  4. Applications of bauxite residue: A mini-review.

    Science.gov (United States)

    Verma, Ajay S; Suri, Narendra M; Kant, Suman

    2017-10-01

    Bauxite residue is the waste generated during alumina production by Bayer's process. The amount of bauxite residue (40-50 wt%) generated depends on the quality of bauxite ore used for the processing. High alkalinity and high caustic content in bauxite residue causes environmental risk for fertile soil and ground water contamination. The caustic (NaOH) content in bauxite residue leads to human health risks, like dermal problems and irritation to eyes. Moreover, disposal of bauxite residue requires a large area; such problems can only be minimised by utilising bauxite residue effectively. For two decades, bauxite residue has been used as a binder in cement industries and filler/reinforcement for composite materials in the automobile industry. Valuable metals and oxides, like alumina (Al 2 O 3 ), titanium oxide (TiO 2 ) and iron oxide Fe 2 O 3 , were extracted from bauxite residue to reduce waste. Bauxite residue was utilised in construction and structure industries to make geopolymers. It was also used in the making of glass-ceramics and a coating material. Recently bauxite residue has been utilised to extract rare earth elements like scandium (Sc), yttrium (Y), lanthanum (La), cerium (Ce), neodymium (Nd) and dysprosium (Dy). In this review article, the mineralogical characteristics of bauxite residue are summarised and current progresses on utilisation of bauxite residue in different fields of science and engineering are presented in detail.

  5. Management of NORM Residues

    International Nuclear Information System (INIS)

    2013-06-01

    The IAEA attaches great importance to the dissemination of information that can assist Member States in the development, implementation, maintenance and continuous improvement of systems, programmes and activities that support the nuclear fuel cycle and nuclear applications, and that address the legacy of past practices and accidents. However, radioactive residues are found not only in nuclear fuel cycle activities, but also in a range of other industrial activities, including: - Mining and milling of metalliferous and non-metallic ores; - Production of non-nuclear fuels, including coal, oil and gas; - Extraction and purification of water (e.g. in the generation of geothermal energy, as drinking and industrial process water; in paper and pulp manufacturing processes); - Production of industrial minerals, including phosphate, clay and building materials; - Use of radionuclides, such as thorium, for properties other than their radioactivity. Naturally occurring radioactive material (NORM) may lead to exposures at some stage of these processes and in the use or reuse of products, residues or wastes. Several IAEA publications address NORM issues with a special focus on some of the more relevant industrial operations. This publication attempts to provide guidance on managing residues arising from different NORM type industries, and on pertinent residue management strategies and technologies, to help Member States gain perspectives on the management of NORM residues

  6. A Hybrid Fuzzy Multi-hop Unequal Clustering Algorithm for Dense Wireless Sensor Networks

    Directory of Open Access Journals (Sweden)

    Shawkat K. Guirguis

    2017-01-01

    Full Text Available Clustering is carried out to explore and solve power dissipation problem in wireless sensor network (WSN. Hierarchical network architecture, based on clustering, can reduce energy consumption, balance traffic load, improve scalability, and prolong network lifetime. However, clustering faces two main challenges: hotspot problem and searching for effective techniques to perform clustering. This paper introduces a fuzzy unequal clustering technique for heterogeneous dense WSNs to determine both final cluster heads and their radii. Proposed fuzzy system blends three effective parameters together which are: the distance to the base station, the density of the cluster, and the deviation of the noders residual energy from the average network energy. Our objectives are achieving gain for network lifetime, energy distribution, and energy consumption. To evaluate the proposed algorithm, WSN clustering based routing algorithms are analyzed, simulated, and compared with obtained results. These protocols are LEACH, SEP, HEED, EEUC, and MOFCA.

  7. Immobilization of acid digestion residue

    International Nuclear Information System (INIS)

    Greenhalgh, W.O.; Allen, C.R.

    1983-01-01

    Acid digestion treatment of nuclear waste is similar to incineration processes and results in the bulk of the waste being reduced in volume and weight to some residual solids termed residue. The residue is composed of various dispersible solid materials and typically contains the resultant radioactivity from the waste. This report describes the immobilization of the residue in portland cement, borosilicate glass, and some other waste forms. Diagrams showing the cement and glass virtification parameters are included in the report as well as process steps and candidate waste product forms. Cement immobilization is simplest and probably least expensive; glass vitrification exhibits the best overall volume reduction ratio

  8. PRESS-based EFOR algorithm for the dynamic parametrical modeling of nonlinear MDOF systems

    Science.gov (United States)

    Liu, Haopeng; Zhu, Yunpeng; Luo, Zhong; Han, Qingkai

    2017-09-01

    In response to the identification problem concerning multi-degree of freedom (MDOF) nonlinear systems, this study presents the extended forward orthogonal regression (EFOR) based on predicted residual sums of squares (PRESS) to construct a nonlinear dynamic parametrical model. The proposed parametrical model is based on the non-linear autoregressive with exogenous inputs (NARX) model and aims to explicitly reveal the physical design parameters of the system. The PRESS-based EFOR algorithm is proposed to identify such a model for MDOF systems. By using the algorithm, we built a common-structured model based on the fundamental concept of evaluating its generalization capability through cross-validation. The resulting model aims to prevent over-fitting with poor generalization performance caused by the average error reduction ratio (AERR)-based EFOR algorithm. Then, a functional relationship is established between the coefficients of the terms and the design parameters of the unified model. Moreover, a 5-DOF nonlinear system is taken as a case to illustrate the modeling of the proposed algorithm. Finally, a dynamic parametrical model of a cantilever beam is constructed from experimental data. Results indicate that the dynamic parametrical model of nonlinear systems, which depends on the PRESS-based EFOR, can accurately predict the output response, thus providing a theoretical basis for the optimal design of modeling methods for MDOF nonlinear systems.

  9. Multi-level iteration optimization for diffusive critical calculation

    International Nuclear Information System (INIS)

    Li Yunzhao; Wu Hongchun; Cao Liangzhi; Zheng Youqi

    2013-01-01

    In nuclear reactor core neutron diffusion calculation, there are usually at least three levels of iterations, namely the fission source iteration, the multi-group scattering source iteration and the within-group iteration. Unnecessary calculations occur if the inner iterations are converged extremely tight. But the convergence of the outer iteration may be affected if the inner ones are converged insufficiently tight. Thus, a common scheme suit for most of the problems was proposed in this work to automatically find the optimized settings. The basic idea is to optimize the relative error tolerance of the inner iteration based on the corresponding convergence rate of the outer iteration. Numerical results of a typical thermal neutron reactor core problem and a fast neutron reactor core problem demonstrate the effectiveness of this algorithm in the variational nodal method code NODAL with the Gauss-Seidel left preconditioned multi-group GMRES algorithm. The multi-level iteration optimization scheme reduces the number of multi-group and within-group iterations respectively by a factor of about 1-2 and 5-21. (authors)

  10. Augmented Lagrangian methods to solve Navier-Stokes equations for a Bingham fluid flow

    International Nuclear Information System (INIS)

    Boscardin, Laetitia

    1999-01-01

    The objective of this research thesis is to develop one or more methods for the numerical resolution of equations of movement obtained for a Bingham fluid. The resolution of Navier-Stokes equations is processed by splitting elliptic and hyperbolic operators (Galerkin transport). In this purpose, the author first studied the Stokes problem, and then addressed issues of stability and consistency of the global scheme. The variational formulation of the Stokes problem can be expressed under the form of a minimisation problem under the constraint of non linear and non differentiable functions. Then, the author proposes a discretization of the Stokes problem based on a hybrid finite element method. Then he extends the demonstrations of stability and consistency of the Galerkin-transport scheme which have been established for a Newtonian fluid, to the case of a Bingham fluid. A relaxation algorithm and a Newton-GMRES algorithm are developed to solve the problem, and their convergence is studied. To ensure this convergence, some constraints must be verified. In order to do so, a specific speed element has been developed [fr

  11. Algorithming the Algorithm

    DEFF Research Database (Denmark)

    Mahnke, Martina; Uprichard, Emma

    2014-01-01

    Imagine sailing across the ocean. The sun is shining, vastness all around you. And suddenly [BOOM] you’ve hit an invisible wall. Welcome to the Truman Show! Ever since Eli Pariser published his thoughts on a potential filter bubble, this movie scenario seems to have become reality, just with slight...... changes: it’s not the ocean, it’s the internet we’re talking about, and it’s not a TV show producer, but algorithms that constitute a sort of invisible wall. Building on this assumption, most research is trying to ‘tame the algorithmic tiger’. While this is a valuable and often inspiring approach, we...

  12. Pseudo-deterministic Algorithms

    OpenAIRE

    Goldwasser , Shafi

    2012-01-01

    International audience; In this talk we describe a new type of probabilistic algorithm which we call Bellagio Algorithms: a randomized algorithm which is guaranteed to run in expected polynomial time, and to produce a correct and unique solution with high probability. These algorithms are pseudo-deterministic: they can not be distinguished from deterministic algorithms in polynomial time by a probabilistic polynomial time observer with black box access to the algorithm. We show a necessary an...

  13. Residual stress characterization of steel TIG welds by neutron diffraction and by residual magnetic stray field mappings

    Science.gov (United States)

    Stegemann, Robert; Cabeza, Sandra; Lyamkin, Viktor; Bruno, Giovanni; Pittner, Andreas; Wimpory, Robert; Boin, Mirko; Kreutzbruck, Marc

    2017-03-01

    The residual stress distribution of tungsten inert gas welded S235JRC+C plates was determined by means of neutron diffraction (ND). Large longitudinal residual stresses with maxima around 600 MPa were found. With these results as reference, the evaluation of residual stress with high spatial resolution GMR (giant magneto resistance) sensors was discussed. The experiments performed indicate a correlation between changes in residual stresses (ND) and the normal component of local residual magnetic stray fields (GMR). Spatial variations in the magnetic field strength perpendicular to the welds are in the order of the magnetic field of the earth.

  14. Radioactive material in residues of health services residues

    International Nuclear Information System (INIS)

    Costa R, A. Jr.; Recio, J.C.

    2006-01-01

    The work presents the operational actions developed by the one organ responsible regulator for the control of the material use radioactive in Brazil. Starting from the appearance of coming radioactive material of hospitals and clinical with services of nuclear medicine, material that that is picked up and transported in specific trucks for the gathering of residuals of hospital origin, and guided one it manufactures of treatment of residuals of services of health, where they suffer radiological monitoring before to guide them for final deposition in sanitary embankment, in the city of Sao Paulo, Brazil. The appearance of this radioactive material exposes a possible one violation of the norms that govern the procedures and practices in that sector in the country. (Author)

  15. The Algorithmic Imaginary

    DEFF Research Database (Denmark)

    Bucher, Taina

    2017-01-01

    the notion of the algorithmic imaginary. It is argued that the algorithmic imaginary – ways of thinking about what algorithms are, what they should be and how they function – is not just productive of different moods and sensations but plays a generative role in moulding the Facebook algorithm itself...... of algorithms affect people's use of these platforms, if at all? To help answer these questions, this article examines people's personal stories about the Facebook algorithm through tweets and interviews with 25 ordinary users. To understand the spaces where people and algorithms meet, this article develops...

  16. Improved crop residue cover estimates by coupling spectral indices for residue and moisture

    Science.gov (United States)

    Remote sensing assessment of soil residue cover (fR) and tillage intensity will improve our predictions of the impact of agricultural practices and promote sustainable management. Spectral indices for estimating fR are sensitive to soil and residue water content, therefore, the uncertainty of estima...

  17. Statistical inference on residual life

    CERN Document Server

    Jeong, Jong-Hyeon

    2014-01-01

    This is a monograph on the concept of residual life, which is an alternative summary measure of time-to-event data, or survival data. The mean residual life has been used for many years under the name of life expectancy, so it is a natural concept for summarizing survival or reliability data. It is also more interpretable than the popular hazard function, especially for communications between patients and physicians regarding the efficacy of a new drug in the medical field. This book reviews existing statistical methods to infer the residual life distribution. The review and comparison includes existing inference methods for mean and median, or quantile, residual life analysis through medical data examples. The concept of the residual life is also extended to competing risks analysis. The targeted audience includes biostatisticians, graduate students, and PhD (bio)statisticians. Knowledge in survival analysis at an introductory graduate level is advisable prior to reading this book.

  18. Efficient predictive algorithms for image compression

    CERN Document Server

    Rosário Lucas, Luís Filipe; Maciel de Faria, Sérgio Manuel; Morais Rodrigues, Nuno Miguel; Liberal Pagliari, Carla

    2017-01-01

    This book discusses efficient prediction techniques for the current state-of-the-art High Efficiency Video Coding (HEVC) standard, focusing on the compression of a wide range of video signals, such as 3D video, Light Fields and natural images. The authors begin with a review of the state-of-the-art predictive coding methods and compression technologies for both 2D and 3D multimedia contents, which provides a good starting point for new researchers in the field of image and video compression. New prediction techniques that go beyond the standardized compression technologies are then presented and discussed. In the context of 3D video, the authors describe a new predictive algorithm for the compression of depth maps, which combines intra-directional prediction, with flexible block partitioning and linear residue fitting. New approaches are described for the compression of Light Field and still images, which enforce sparsity constraints on linear models. The Locally Linear Embedding-based prediction method is in...

  19. An Adaptive Filtering Algorithm Based on Genetic Algorithm-Backpropagation Network

    Directory of Open Access Journals (Sweden)

    Kai Hu

    2013-01-01

    Full Text Available A new image filtering algorithm is proposed. GA-BPN algorithm uses genetic algorithm (GA to decide weights in a back propagation neural network (BPN. It has better global optimal characteristics than traditional optimal algorithm. In this paper, we used GA-BPN to do image noise filter researching work. Firstly, this paper uses training samples to train GA-BPN as the noise detector. Then, we utilize the well-trained GA-BPN to recognize noise pixels in target image. And at last, an adaptive weighted average algorithm is used to recover noise pixels recognized by GA-BPN. Experiment data shows that this algorithm has better performance than other filters.

  20. Fuzzy-Logic Based Distributed Energy-Efficient Clustering Algorithm for Wireless Sensor Networks.

    Science.gov (United States)

    Zhang, Ying; Wang, Jun; Han, Dezhi; Wu, Huafeng; Zhou, Rundong

    2017-07-03

    Due to the high-energy efficiency and scalability, the clustering routing algorithm has been widely used in wireless sensor networks (WSNs). In order to gather information more efficiently, each sensor node transmits data to its Cluster Head (CH) to which it belongs, by multi-hop communication. However, the multi-hop communication in the cluster brings the problem of excessive energy consumption of the relay nodes which are closer to the CH. These nodes' energy will be consumed more quickly than the farther nodes, which brings the negative influence on load balance for the whole networks. Therefore, we propose an energy-efficient distributed clustering algorithm based on fuzzy approach with non-uniform distribution (EEDCF). During CHs' election, we take nodes' energies, nodes' degree and neighbor nodes' residual energies into consideration as the input parameters. In addition, we take advantage of Takagi, Sugeno and Kang (TSK) fuzzy model instead of traditional method as our inference system to guarantee the quantitative analysis more reasonable. In our scheme, each sensor node calculates the probability of being as CH with the help of fuzzy inference system in a distributed way. The experimental results indicate EEDCF algorithm is better than some current representative methods in aspects of data transmission, energy consumption and lifetime of networks.

  1. Iterative local Chi2 alignment algorithm for the ATLAS Pixel detector

    CERN Document Server

    Göttfert, Tobias

    The existing local chi2 alignment approach for the ATLAS SCT detector was extended to the alignment of the ATLAS Pixel detector. This approach is linear, aligns modules separately, and uses distance of closest approach residuals and iterations. The derivation and underlying concepts of the approach are presented. To show the feasibility of the approach for Pixel modules, a simplified, stand-alone track simulation, together with the alignment algorithm, was developed with the ROOT analysis software package. The Pixel alignment software was integrated into Athena, the ATLAS software framework. First results and the achievable accuracy for this approach with a simulated dataset are presented.

  2. Nature-inspired optimization algorithms

    CERN Document Server

    Yang, Xin-She

    2014-01-01

    Nature-Inspired Optimization Algorithms provides a systematic introduction to all major nature-inspired algorithms for optimization. The book's unified approach, balancing algorithm introduction, theoretical background and practical implementation, complements extensive literature with well-chosen case studies to illustrate how these algorithms work. Topics include particle swarm optimization, ant and bee algorithms, simulated annealing, cuckoo search, firefly algorithm, bat algorithm, flower algorithm, harmony search, algorithm analysis, constraint handling, hybrid methods, parameter tuning

  3. Residual Structures in Latent Growth Curve Modeling

    Science.gov (United States)

    Grimm, Kevin J.; Widaman, Keith F.

    2010-01-01

    Several alternatives are available for specifying the residual structure in latent growth curve modeling. Two specifications involve uncorrelated residuals and represent the most commonly used residual structures. The first, building on repeated measures analysis of variance and common specifications in multilevel models, forces residual variances…

  4. Convex hull ranking algorithm for multi-objective evolutionary algorithms

    NARCIS (Netherlands)

    Davoodi Monfrared, M.; Mohades, A.; Rezaei, J.

    2012-01-01

    Due to many applications of multi-objective evolutionary algorithms in real world optimization problems, several studies have been done to improve these algorithms in recent years. Since most multi-objective evolutionary algorithms are based on the non-dominated principle, and their complexity

  5. Total algorithms

    NARCIS (Netherlands)

    Tel, G.

    We define the notion of total algorithms for networks of processes. A total algorithm enforces that a "decision" is taken by a subset of the processes, and that participation of all processes is required to reach this decision. Total algorithms are an important building block in the design of

  6. Computing Decoupled Residuals for Compact Disc Players

    DEFF Research Database (Denmark)

    Odgaard, Peter Fogh; Stoustrup, Jakob; Andersen, Palle

    2006-01-01

    a pair of residuals generated by Compact Disc Player. However, these residuals depend on the performance of position servos in the Compact Disc Player. In other publications of the same authors a pair of decoupled residuals is derived. However, the computation of these alternative residuals has been...

  7. Characterisation and management of concrete grinding residuals.

    Science.gov (United States)

    Kluge, Matt; Gupta, Nautasha; Watts, Ben; Chadik, Paul A; Ferraro, Christopher; Townsend, Timothy G

    2018-02-01

    Concrete grinding residue is the waste product resulting from the grinding, cutting, and resurfacing of concrete pavement. Potential beneficial applications for concrete grinding residue include use as a soil amendment and as a construction material, including as an additive to Portland cement concrete. Concrete grinding residue exhibits a high pH, and though not hazardous, it is sufficiently elevated that precautions need to be taken around aquatic ecosystems. Best management practices and state regulations focus on reducing the impact on such aquatic environment. Heavy metals are present in concrete grinding residue, but concentrations are of the same magnitude as typically recycled concrete residuals. The chemical composition of concrete grinding residue makes it a useful product for some soil amendment purposes at appropriate land application rates. The presence of unreacted concrete in concrete grinding residue was examined for potential use as partial replacement of cement in new concrete. Testing of Florida concrete grinding residue revealed no dramatic reactivity or improvement in mortar strength.

  8. Progressive geometric algorithms

    NARCIS (Netherlands)

    Alewijnse, S.P.A.; Bagautdinov, T.M.; de Berg, M.T.; Bouts, Q.W.; ten Brink, Alex P.; Buchin, K.A.; Westenberg, M.A.

    2015-01-01

    Progressive algorithms are algorithms that, on the way to computing a complete solution to the problem at hand, output intermediate solutions that approximate the complete solution increasingly well. We present a framework for analyzing such algorithms, and develop efficient progressive algorithms

  9. Progressive geometric algorithms

    NARCIS (Netherlands)

    Alewijnse, S.P.A.; Bagautdinov, T.M.; Berg, de M.T.; Bouts, Q.W.; Brink, ten A.P.; Buchin, K.; Westenberg, M.A.

    2014-01-01

    Progressive algorithms are algorithms that, on the way to computing a complete solution to the problem at hand, output intermediate solutions that approximate the complete solution increasingly well. We present a framework for analyzing such algorithms, and develop efficient progressive algorithms

  10. Residual symptoms and functioning in depression, does the type of residual symptom matter? A post-hoc analysis

    Directory of Open Access Journals (Sweden)

    Romera Irene

    2013-02-01

    Full Text Available Abstract Background The degrees to which residual symptoms in major depressive disorder (MDD adversely affect patient functioning is not known. This post-hoc analysis explored the association between different residual symptoms and patient functioning. Methods Patients with MDD who responded (≥50% on the 17-item Hamilton Rating Scale for Depression; HAMD-17 after 3 months of treatment (624/930 were included. Residual core mood-symptoms (HAMD-17 core symptom subscale ≥1, residual insomnia-symptoms (HAMD-17 sleep subscale ≥1, residual anxiety-symptoms (HAMD-17-anxiety subscale ≥1, residual somatic-symptoms (HAMD-17 Item 13 ≥1, pain (Visual Analogue Scale ≥30, and functioning were assessed after 3 months treatment. A stepwise logistic regression model with normal functioning (Social and Occupational Functioning Assessment Scale ≥80 as the dependent variable was used. Results After 3 months, 59.5% of patients (371/624 achieved normal functioning and 66.0% (412/624 were in remission. Residual symptom prevalence was: core mood symptoms 72%; insomnia 63%; anxiety 78%; and somatic symptoms 41%. Pain reported in 18%. Factors associated with normal functioning were absence of core mood symptoms (odds ratio [OR] 8.7; 95% confidence interval [CI], 4.6–16.7, absence of insomnia symptoms (OR 1.8; 95% CI, 1.2–2.7, episode length (4–24 weeks vs. ≥24 weeks [OR 2.0; 95% CI, 1.1–3.6] and better baseline functioning (OR 1.0; 95% CI, 1.0–1.1. A significant interaction between residual anxiety symptoms and pain was found (p = 0.0080. Conclusions Different residual symptoms are associated to different degrees with patient functioning. To achieve normal functioning, specific residual symptoms domains might be targeted for treatment.

  11. Nitrogen availability of biogas residues

    Energy Technology Data Exchange (ETDEWEB)

    El-Sayed Fouda, Sara

    2011-09-07

    The objectives of this study were to characterize biogas residues either unseparated or separated into a liquid and a solid phase from the fermentation of different substrates with respect to their N and C content. In addition, short and long term effects of the application of these biogas residues on the N availability and N utilization by ryegrass was investigated. It is concluded that unseparated or liquid separated biogas residues provide N at least corresponding to their ammonium content and that after the first fertilizer application the C{sub org}:N{sub org} ratio of the biogas residues was a crucial factor for the N availability. After long term application, the organic N accumulated in the soil leads to an increased release of N.

  12. Archimedean copula estimation of distribution algorithm based on artificial bee colony algorithm

    Institute of Scientific and Technical Information of China (English)

    Haidong Xu; Mingyan Jiang; Kun Xu

    2015-01-01

    The artificial bee colony (ABC) algorithm is a com-petitive stochastic population-based optimization algorithm. How-ever, the ABC algorithm does not use the social information and lacks the knowledge of the problem structure, which leads to in-sufficiency in both convergent speed and searching precision. Archimedean copula estimation of distribution algorithm (ACEDA) is a relatively simple, time-economic and multivariate correlated EDA. This paper proposes a novel hybrid algorithm based on the ABC algorithm and ACEDA cal ed Archimedean copula estima-tion of distribution based on the artificial bee colony (ACABC) algorithm. The hybrid algorithm utilizes ACEDA to estimate the distribution model and then uses the information to help artificial bees to search more efficiently in the search space. Six bench-mark functions are introduced to assess the performance of the ACABC algorithm on numerical function optimization. Experimen-tal results show that the ACABC algorithm converges much faster with greater precision compared with the ABC algorithm, ACEDA and the global best (gbest)-guided ABC (GABC) algorithm in most of the experiments.

  13. Residual stresses around Vickers indents

    International Nuclear Information System (INIS)

    Pajares, A.; Guiberteau, F.; Steinbrech, R.W.

    1995-01-01

    The residual stresses generated by Vickers indentation in brittle materials and their changes due to annealing and surface removal were studied in 4 mol% yttria partially stabilized zirconia (4Y-PSZ). Three experimental methods to gain information about the residual stress field were applied: (i) crack profile measurements based on serial sectioning, (ii) controlled crack propagation in post indentation bending tests and (iii) double indentation tests with smaller secondary indents located around a larger primary impression. Three zones of different residual stress behavior are deduced from the experiments. Beneath the impression a crack free spherical zone of high hydrostatic stresses exists. This core zone is followed by a transition regime where indentation cracks develop but still experience hydrostatic stresses. Finally, in an outward third zone, the crack contour is entirely governed by the tensile residual stress intensity (elastically deformed region). Annealing and surface removal reduce this crack driving stress intensity. The specific changes of the residual stresses due to the post indentation treatments are described and discussed in detail for the three zones

  14. Minimization of zirconium chlorinator residues

    International Nuclear Information System (INIS)

    Green, G.K.; Harbuck, D.D.

    1995-01-01

    Zirconium chlorinator residues contain an array of rare earths, scandium, unreacted coke, and radioactive thorium and radium. Because of the radioactivity, the residues must be disposed in special waste containment facilities. As these sites become more congested, and with stricter environmental regulations, disposal of large volumes of wastes may become more difficult. To reduce the mass of disposed material, the US Bureau of Mines (USBM) developed technology to recover rare earths, thorium and radium, and unreacted coke from these residues. This technology employs an HCl leach to solubilize over 99% of the scandium and thorium, and over 90% of the rare earths. The leach liquor is processed through several solvent extraction stages to selectively recover scandium, thorium, and rare earths. The leach residue is further leached with an organic acid to solubilize radium, thus allowing unreacted coke to be recycled to the chlorinator. The thorium and radium waste products, which comprise only 2.1% of the original residue mass, can then be sent to the radioactive waste facility

  15. 9 CFR 311.39 - Biological residues.

    Science.gov (United States)

    2010-01-01

    ... 9 Animals and Animal Products 2 2010-01-01 2010-01-01 false Biological residues. 311.39 Section... Biological residues. Carcasses, organs, or other parts of carcasses of livestock shall be condemned if it is determined that they are adulterated because of the presence of any biological residues. ...

  16. A filtered backprojection algorithm with characteristics of the iterative landweber algorithm

    OpenAIRE

    L. Zeng, Gengsheng

    2012-01-01

    Purpose: In order to eventually develop an analytical algorithm with noise characteristics of an iterative algorithm, this technical note develops a window function for the filtered backprojection (FBP) algorithm in tomography that behaves as an iterative Landweber algorithm.

  17. Multi-rate cubature Kalman filter based data fusion method with residual compensation to adapt to sampling rate discrepancy in attitude measurement system.

    Science.gov (United States)

    Guo, Xiaoting; Sun, Changku; Wang, Peng

    2017-08-01

    This paper investigates the multi-rate inertial and vision data fusion problem in nonlinear attitude measurement systems, where the sampling rate of the inertial sensor is much faster than that of the vision sensor. To fully exploit the high frequency inertial data and obtain favorable fusion results, a multi-rate CKF (Cubature Kalman Filter) algorithm with estimated residual compensation is proposed in order to adapt to the problem of sampling rate discrepancy. During inter-sampling of slow observation data, observation noise can be regarded as infinite. The Kalman gain is unknown and approaches zero. The residual is also unknown. Therefore, the filter estimated state cannot be compensated. To obtain compensation at these moments, state error and residual formulas are modified when compared with the observation data available moments. Self-propagation equation of the state error is established to propagate the quantity from the moments with observation to the moments without observation. Besides, a multiplicative adjustment factor is introduced as Kalman gain, which acts on the residual. Then the filter estimated state can be compensated even when there are no visual observation data. The proposed method is tested and verified in a practical setup. Compared with multi-rate CKF without residual compensation and single-rate CKF, a significant improvement is obtained on attitude measurement by using the proposed multi-rate CKF with inter-sampling residual compensation. The experiment results with superior precision and reliability show the effectiveness of the proposed method.

  18. Residual-stress measurements

    Energy Technology Data Exchange (ETDEWEB)

    Ezeilo, A N; Webster, G A [Imperial College, London (United Kingdom); Webster, P J [Salford Univ. (United Kingdom)

    1997-04-01

    Because neutrons can penetrate distances of up to 50 mm in most engineering materials, this makes them unique for establishing residual-stress distributions non-destructively. D1A is particularly suited for through-surface measurements as it does not suffer from instrumental surface aberrations commonly found on multidetector instruments, while D20 is best for fast internal-strain scanning. Two examples for residual-stress measurements in a shot-peened material, and in a weld are presented to demonstrate the attractive features of both instruments. (author).

  19. Residues from waste incineration. Final report

    Energy Technology Data Exchange (ETDEWEB)

    Astrup, T.; Juul Pedersen, A.; Hyks, J.; Frandsen, F.J.

    2009-08-15

    The overall objective of the project was to improve the understanding of the formation and characteristics of residues from waste incineration. This was done focusing on the importance of the waste input and the operational conditions of the furnace. Data and results obtained from the project have been discussed in this report according to the following three overall parts: i) mass flows and element distribution, ii) flue gas/particle partitioning and corrosion/deposition aspects, and iii) residue leaching. This has been done with the intent of structuring the discussion while tacitly acknowledging that these aspects are interrelated and cannot be separated. Overall, it was found that the waste input composition had significant impact of the characteristics of the generated residues. A similar correlation between operational conditions and residue characteristics could not be observed. Consequently, the project recommend that optimization of residue quality should focus on controlling the waste input composition. The project results showed that including specific waste materials (and thereby also excluding the same materials) may have significant effects on the residue composition, residue leaching, aerosol and deposit formation.It is specifically recommended to minimize Cl in the input waste. Based on the project results, it was found that a significant potential for optimization of waste incineration exist. (author)

  20. Super-Encryption Implementation Using Monoalphabetic Algorithm and XOR Algorithm for Data Security

    Science.gov (United States)

    Rachmawati, Dian; Andri Budiman, Mohammad; Aulia, Indra

    2018-03-01

    The exchange of data that occurs offline and online is very vulnerable to the threat of data theft. In general, cryptography is a science and art to maintain data secrecy. An encryption is a cryptography algorithm in which data is transformed into cipher text, which is something that is unreadable and meaningless so it cannot be read or understood by other parties. In super-encryption, two or more encryption algorithms are combined to make it more secure. In this work, Monoalphabetic algorithm and XOR algorithm are combined to form a super- encryption. Monoalphabetic algorithm works by changing a particular letter into a new letter based on existing keywords while the XOR algorithm works by using logic operation XOR Since Monoalphabetic algorithm is a classical cryptographic algorithm and XOR algorithm is a modern cryptographic algorithm, this scheme is expected to be both easy-to-implement and more secure. The combination of the two algorithms is capable of securing the data and restoring it back to its original form (plaintext), so the data integrity is still ensured.

  1. A Novel Multiscale Ensemble Carbon Price Prediction Model Integrating Empirical Mode Decomposition, Genetic Algorithm and Artificial Neural Network

    Directory of Open Access Journals (Sweden)

    Bangzhu Zhu

    2012-02-01

    Full Text Available Due to the movement and complexity of the carbon market, traditional monoscale forecasting approaches often fail to capture its nonstationary and nonlinear properties and accurately describe its moving tendencies. In this study, a multiscale ensemble forecasting model integrating empirical mode decomposition (EMD, genetic algorithm (GA and artificial neural network (ANN is proposed to forecast carbon price. Firstly, the proposed model uses EMD to decompose carbon price data into several intrinsic mode functions (IMFs and one residue. Then, the IMFs and residue are composed into a high frequency component, a low frequency component and a trend component which have similar frequency characteristics, simple components and strong regularity using the fine-to-coarse reconstruction algorithm. Finally, those three components are predicted using an ANN trained by GA, i.e., a GAANN model, and the final forecasting results can be obtained by the sum of these three forecasting results. For verification and testing, two main carbon future prices with different maturity in the European Climate Exchange (ECX are used to test the effectiveness of the proposed multiscale ensemble forecasting model. Empirical results obtained demonstrate that the proposed multiscale ensemble forecasting model can outperform the single random walk (RW, ARIMA, ANN and GAANN models without EMD preprocessing and the ensemble ARIMA model with EMD preprocessing.

  2. Ensemble Kalman filtering with residual nudging

    KAUST Repository

    Luo, X.

    2012-10-03

    Covariance inflation and localisation are two important techniques that are used to improve the performance of the ensemble Kalman filter (EnKF) by (in effect) adjusting the sample covariances of the estimates in the state space. In this work, an additional auxiliary technique, called residual nudging, is proposed to monitor and, if necessary, adjust the residual norms of state estimates in the observation space. In an EnKF with residual nudging, if the residual norm of an analysis is larger than a pre-specified value, then the analysis is replaced by a new one whose residual norm is no larger than a pre-specified value. Otherwise, the analysis is considered as a reasonable estimate and no change is made. A rule for choosing the pre-specified value is suggested. Based on this rule, the corresponding new state estimates are explicitly derived in case of linear observations. Numerical experiments in the 40-dimensional Lorenz 96 model show that introducing residual nudging to an EnKF may improve its accuracy and/or enhance its stability against filter divergence, especially in the small ensemble scenario.

  3. Development of computer program ENMASK for prediction of residual environmental masking-noise spectra, from any three independent environmental parameters

    Energy Technology Data Exchange (ETDEWEB)

    Chang, Y.-S.; Liebich, R. E.; Chun, K. C.

    2000-03-31

    Residual environmental sound can mask intrusive4 (unwanted) sound. It is a factor that can affect noise impacts and must be considered both in noise-impact studies and in noise-mitigation designs. Models for quantitative prediction of sensation level (audibility) and psychological effects of intrusive noise require an input with 1/3 octave-band spectral resolution of environmental masking noise. However, the majority of published residual environmental masking-noise data are given with either octave-band frequency resolution or only single A-weighted decibel values. A model has been developed that enables estimation of 1/3 octave-band residual environmental masking-noise spectra and relates certain environmental parameters to A-weighted sound level. This model provides a correlation among three environmental conditions: measured residual A-weighted sound-pressure level, proximity to a major roadway, and population density. Cited field-study data were used to compute the most probable 1/3 octave-band sound-pressure spectrum corresponding to any selected one of these three inputs. In turn, such spectra can be used as an input to models for prediction of noise impacts. This paper discusses specific algorithms included in the newly developed computer program ENMASK. In addition, the relative audibility of the environmental masking-noise spectra at different A-weighted sound levels is discussed, which is determined by using the methodology of program ENAUDIBL.

  4. Determination of Pesticides Residues in Cucumbers Grown in Greenhouse and the Effect of Some Procedures on Their Residues.

    Science.gov (United States)

    Leili, Mostafa; Pirmoghani, Amin; Samadi, Mohammad Taghi; Shokoohi, Reza; Roshanaei, Ghodratollah; Poormohammadi, Ali

    2016-11-01

    The objective of this study was to determine the residual concentrations of ethion and imidacloprid in cucumbers grown in greenhouse. The effect of some simple processing procedures on both ethion and imidacloprid residues were also studied. Ten active greenhouses that produce cucumber were randomly selected. Ethion and imidacloprid as the most widely used pesticides were measured in cucumber samples of studied greenhouses. Moreover, the effect of storing, washing, and peeling as simple processing procedures on both ethion and imidacloprid residues were investigated. One hour after pesticide application; the maximum residue levels (MRLs) of ethion and imidacloprid were higher than that of Codex standard level. One day after pesticide application, the levels of pesticides were decreased about 35 and 31% for ethion and imidacloprid, respectively, which still were higher than the MRL. Washing procedure led to about 51 and 42.5% loss in ethion and imidacloprid residues, respectively. Peeling procedure also led to highest loss of 93.4 and 63.7% in ethion and imidacloprid residues, respectively. The recovery for both target analytes was in the range between 88 and 102%. The residue values in collected samples one hour after pesticides application were higher than standard value. The storing, washing, and peeling procedures lead to the decrease of pesticide residues in greenhouse cucumbers. Among them, the peeling procedure has the greatest impact on residual reduction. Therefore, these procedures can be used as simple and effective processing techniques for reducing and removing pesticides from greenhouse products before their consumption.

  5. Alternatives to crop residues for soil amendment

    OpenAIRE

    Powell, J.M.; Unger, P.W.

    1997-01-01

    Metadata only record In semiarid agroecosystems, crop residues can provide important benefits of soil and water conservation, nutrient cycling, and improved subsequent crop yields. However, there are frequently multiple competing uses for residues, including animal forage, fuel, and construction material. This chapter discusses the various uses of crop residues and examines alternative soil amendments when crop residues cannot be left on the soil.

  6. 77 FR 24671 - Compliance Guide for Residue Prevention and Agency Testing Policy for Residues

    Science.gov (United States)

    2012-04-25

    ... Hazard Analysis and Critical Control Points (HACCP) inspection system, another important component of the NRP is to provide verification of residue control in HACCP systems. As part of the HACCP regulation... guide, and FSIS finds violative residues, the establishment's HACCP system may be inadequate under 9 CFR...

  7. Linear feature detection algorithm for astronomical surveys - I. Algorithm description

    Science.gov (United States)

    Bektešević, Dino; Vinković, Dejan

    2017-11-01

    Computer vision algorithms are powerful tools in astronomical image analyses, especially when automation of object detection and extraction is required. Modern object detection algorithms in astronomy are oriented towards detection of stars and galaxies, ignoring completely the detection of existing linear features. With the emergence of wide-field sky surveys, linear features attract scientific interest as possible trails of fast flybys of near-Earth asteroids and meteors. In this work, we describe a new linear feature detection algorithm designed specifically for implementation in big data astronomy. The algorithm combines a series of algorithmic steps that first remove other objects (stars and galaxies) from the image and then enhance the line to enable more efficient line detection with the Hough algorithm. The rate of false positives is greatly reduced thanks to a step that replaces possible line segments with rectangles and then compares lines fitted to the rectangles with the lines obtained directly from the image. The speed of the algorithm and its applicability in astronomical surveys are also discussed.

  8. Golden Sine Algorithm: A Novel Math-Inspired Algorithm

    Directory of Open Access Journals (Sweden)

    TANYILDIZI, E.

    2017-05-01

    Full Text Available In this study, Golden Sine Algorithm (Gold-SA is presented as a new metaheuristic method for solving optimization problems. Gold-SA has been developed as a new search algorithm based on population. This math-based algorithm is inspired by sine that is a trigonometric function. In the algorithm, random individuals are created as many as the number of search agents with uniform distribution for each dimension. The Gold-SA operator searches to achieve a better solution in each iteration by trying to bring the current situation closer to the target value. The solution space is narrowed by the golden section so that the areas that are supposed to give only good results are scanned instead of the whole solution space scan. In the tests performed, it is seen that Gold-SA has better results than other population based methods. In addition, Gold-SA has fewer algorithm-dependent parameters and operators than other metaheuristic methods, increasing the importance of this method by providing faster convergence of this new method.

  9. Computational models for residual creep life prediction of power plant components

    International Nuclear Information System (INIS)

    Grewal, G.S.; Singh, A.K.; Ramamoortry, M.

    2006-01-01

    All high temperature - high pressure power plant components are prone to irreversible visco-plastic deformation by the phenomenon of creep. The steady state creep response as well as the total creep life of a material is related to the operational component temperature through, respectively, the exponential and inverse exponential relationships. Minor increases in the component temperature can thus have serious consequences as far as the creep life and dimensional stability of a plant component are concerned. In high temperature steam tubing in power plants, one mechanism by which a significant temperature rise can occur is by the growth of a thermally insulating oxide film on its steam side surface. In the present paper, an elegantly simple and computationally efficient technique is presented for predicting the residual creep life of steel components subjected to continual steam side oxide film growth. Similarly, fabrication of high temperature power plant components involves extensive use of welding as the fabrication process of choice. Naturally, issues related to the creep life of weldments have to be seriously addressed for safe and continual operation of the welded plant component. Unfortunately, a typical weldment in an engineering structure is a zone of complex microstructural gradation comprising of a number of distinct sub-zones with distinct meso-scale and micro-scale morphology of the phases and (even) chemistry and its creep life prediction presents considerable challenges. The present paper presents a stochastic algorithm, which can be' used for developing experimental creep-cavitation intensity versus residual life correlations for welded structures. Apart from estimates of the residual life in a mean field sense, the model can be used for predicting the reliability of the plant component in a rigorous probabilistic setting. (author)

  10. The Orthogonally Partitioned EM Algorithm: Extending the EM Algorithm for Algorithmic Stability and Bias Correction Due to Imperfect Data.

    Science.gov (United States)

    Regier, Michael D; Moodie, Erica E M

    2016-05-01

    We propose an extension of the EM algorithm that exploits the common assumption of unique parameterization, corrects for biases due to missing data and measurement error, converges for the specified model when standard implementation of the EM algorithm has a low probability of convergence, and reduces a potentially complex algorithm into a sequence of smaller, simpler, self-contained EM algorithms. We use the theory surrounding the EM algorithm to derive the theoretical results of our proposal, showing that an optimal solution over the parameter space is obtained. A simulation study is used to explore the finite sample properties of the proposed extension when there is missing data and measurement error. We observe that partitioning the EM algorithm into simpler steps may provide better bias reduction in the estimation of model parameters. The ability to breakdown a complicated problem in to a series of simpler, more accessible problems will permit a broader implementation of the EM algorithm, permit the use of software packages that now implement and/or automate the EM algorithm, and make the EM algorithm more accessible to a wider and more general audience.

  11. Parallel sorting algorithms

    CERN Document Server

    Akl, Selim G

    1985-01-01

    Parallel Sorting Algorithms explains how to use parallel algorithms to sort a sequence of items on a variety of parallel computers. The book reviews the sorting problem, the parallel models of computation, parallel algorithms, and the lower bounds on the parallel sorting problems. The text also presents twenty different algorithms, such as linear arrays, mesh-connected computers, cube-connected computers. Another example where algorithm can be applied is on the shared-memory SIMD (single instruction stream multiple data stream) computers in which the whole sequence to be sorted can fit in the

  12. The Non–Symmetric s–Step Lanczos Algorithm: Derivation of Efficient Recurrences and Synchronization–Reducing Variants of BiCG and QMR

    Directory of Open Access Journals (Sweden)

    Feuerriegel Stefan

    2015-12-01

    Full Text Available The Lanczos algorithm is among the most frequently used iterative techniques for computing a few dominant eigenvalues of a large sparse non-symmetric matrix. At the same time, it serves as a building block within biconjugate gradient (BiCG and quasi-minimal residual (QMR methods for solving large sparse non-symmetric systems of linear equations. It is well known that, when implemented on distributed-memory computers with a huge number of processes, the synchronization time spent on computing dot products increasingly limits the parallel scalability. Therefore, we propose synchronization-reducing variants of the Lanczos, as well as BiCG and QMR methods, in an attempt to mitigate these negative performance effects. These so-called s-step algorithms are based on grouping dot products for joint execution and replacing time-consuming matrix operations by efficient vector recurrences. The purpose of this paper is to provide a rigorous derivation of the recurrences for the s-step Lanczos algorithm, introduce s-step BiCG and QMR variants, and compare the parallel performance of these new s-step versions with previous algorithms.

  13. Cycling of grain legume residue nitrogen

    DEFF Research Database (Denmark)

    Jensen, E.S.

    1995-01-01

    Symbiotic nitrogen fixation by legumes is the main input of nitrogen in ecological agriculture. The cycling of N-15-labelled mature pea (Pisum sativum L.) residues was studied during three years in small field plots and lysimeters. The residual organic labelled N declined rapidly during the initial...... management methods in order to conserve grain legume residue N sources within the soil-plant system....

  14. On root class residuality of HNN-extensions

    International Nuclear Information System (INIS)

    Tieudjo, D.

    2004-08-01

    A sufficient condition or root-class residuality of HNN-extensions with root-class residual base group is proven; namely if G = -1 1Ht = K, φ> is the HNN-extension with base group A, stable letter t and associated subgroups H and K via the isomorphism φ, then G is root-class residual if group A is root-class residual and there exists a homomorphism σ of group G onto some group of a root-class such that σ is one-to-one on H. For the particular case when H = K and σ is the identical map, it is shown that G is root-class residual if and only if A is root-class residual and subgroup H of A is root-class separable. These results are generalized to multiple HNN-extensions. (author)

  15. Distribution of Penicillin G Residues in Culled Dairy Cow Muscles: Implications for Residue Monitoring

    Science.gov (United States)

    The U.S. Food and Drug Administration sets tolerances for veterinary drug residues in muscle, but does not specify which type of muscle should be analyzed. In order to determine if antibiotic residue levels are dependent on muscle type, 7 culled dairy cows were dosed with Penicillin G (Pen G) from ...

  16. Site-Mutation of Hydrophobic Core Residues Synchronically Poise Super Interleukin 2 for Signaling: Identifying Distant Structural Effects through Affordable Computations

    Directory of Open Access Journals (Sweden)

    Longcan Mei

    2018-03-01

    Full Text Available A superkine variant of interleukin-2 with six site mutations away from the binding interface developed from the yeast display technique has been previously characterized as undergoing a distal structure alteration which is responsible for its super-potency and provides an elegant case study with which to get insight about how to utilize allosteric effect to achieve desirable protein functions. By examining the dynamic network and the allosteric pathways related to those mutated residues using various computational approaches, we found that nanosecond time scale all-atom molecular dynamics simulations can identify the dynamic network as efficient as an ensemble algorithm. The differentiated pathways for the six core residues form a dynamic network that outlines the area of structure alteration. The results offer potentials of using affordable computing power to predict allosteric structure of mutants in knowledge-based mutagenesis.

  17. Environmental dredging residual generation and management.

    Science.gov (United States)

    Patmont, Clay; LaRosa, Paul; Narayanan, Raghav; Forrest, Casey

    2018-05-01

    The presence and magnitude of sediment contamination remaining in a completed dredge area can often dictate the success of an environmental dredging project. The need to better understand and manage this remaining contamination, referred to as "postdredging residuals," has increasingly been recognized by practitioners and investigators. Based on recent dredging projects with robust characterization programs, it is now understood that the residual contamination layer in the postdredging sediment comprises a mixture of contaminated sediments that originate from throughout the dredge cut. This mixture of contaminated sediments initially exhibits fluid mud properties that can contribute to sediment transport and contamination risk outside of the dredge area. This article reviews robust dredging residual evaluations recently performed in the United States and Canada, including the Hudson River, Lower Fox River, Ashtabula River, and Esquimalt Harbour, along with other projects. These data better inform the understanding of residuals generation, leading to improved models of dredging residual formation to inform remedy evaluation, selection, design, and implementation. Data from these projects confirm that the magnitude of dredging residuals is largely determined by site conditions, primarily in situ sediment fluidity or liquidity as measured by dry bulk density. While the generation of dredging residuals cannot be avoided, residuals can be successfully and efficiently managed through careful development and implementation of site-specific management plans. Integr Environ Assess Manag 2018;14:335-343. © 2018 The Authors. Integrated Environmental Assessment and Management Published by Wiley Periodicals, Inc. on behalf of Society of Environmental Toxicology & Chemistry (SETAC). © 2018 The Authors. Integrated Environmental Assessment and Management Published by Wiley Periodicals, Inc. on behalf of Society of Environmental Toxicology & Chemistry (SETAC).

  18. New sparse matrix solver in the KIKO3D 3-dimensional reactor dynamics code

    International Nuclear Information System (INIS)

    Panka, I.; Kereszturi, A.; Hegedus, C.

    2005-01-01

    The goal of this paper is to present a more effective method Bi-CGSTAB for accelerating the large sparse matrix equation solution in the KIKO3D code. This equation system is obtained by using the factorization of the improved quasi static (IQS) method for the time dependent nodal kinetic equations. In the old methodology standard large sparse matrix techniques were considered, where Gauss-Seidel preconditioning and a GMRES-type solver were applied. The validation of KIKO3D using Bi-CGSTAB has been performed by solving of a VVER-1000 kinetic benchmark problem. Additionally, the convergence characteristics were investigated in given macro time steps of Control Rod Ejection transients. The results have been obtained by the old GMRES and new Bi-CGSTAB methods are compared. (author)

  19. Ensemble Kalman filtering with residual nudging

    Directory of Open Access Journals (Sweden)

    Xiaodong Luo

    2012-10-01

    Full Text Available Covariance inflation and localisation are two important techniques that are used to improve the performance of the ensemble Kalman filter (EnKF by (in effect adjusting the sample covariances of the estimates in the state space. In this work, an additional auxiliary technique, called residual nudging, is proposed to monitor and, if necessary, adjust the residual norms of state estimates in the observation space. In an EnKF with residual nudging, if the residual norm of an analysis is larger than a pre-specified value, then the analysis is replaced by a new one whose residual norm is no larger than a pre-specified value. Otherwise, the analysis is considered as a reasonable estimate and no change is made. A rule for choosing the pre-specified value is suggested. Based on this rule, the corresponding new state estimates are explicitly derived in case of linear observations. Numerical experiments in the 40-dimensional Lorenz 96 model show that introducing residual nudging to an EnKF may improve its accuracy and/or enhance its stability against filter divergence, especially in the small ensemble scenario.

  20. Computational Prediction of Hot Spot Residues

    Science.gov (United States)

    Morrow, John Kenneth; Zhang, Shuxing

    2013-01-01

    Most biological processes involve multiple proteins interacting with each other. It has been recently discovered that certain residues in these protein-protein interactions, which are called hot spots, contribute more significantly to binding affinity than others. Hot spot residues have unique and diverse energetic properties that make them challenging yet important targets in the modulation of protein-protein complexes. Design of therapeutic agents that interact with hot spot residues has proven to be a valid methodology in disrupting unwanted protein-protein interactions. Using biological methods to determine which residues are hot spots can be costly and time consuming. Recent advances in computational approaches to predict hot spots have incorporated a myriad of features, and have shown increasing predictive successes. Here we review the state of knowledge around protein-protein interactions, hot spots, and give an overview of multiple in silico prediction techniques of hot spot residues. PMID:22316154

  1. Thermal residual stresses in amorphous thermoplastic polymers

    Science.gov (United States)

    Grassia, Luigi; D'Amore, Alberto

    2010-06-01

    An attempt to calculate the internal stresses in a cylindrically shaped polycarbonate (LEXAN-GE) component, subjected to an arbitrary cooling rate, will be described. The differential volume relaxation arising as a result of the different thermal history suffered by each body point was considered as the primary source of stresses build up [1-3]. A numerical routine was developed accounting for the simultaneous stress and structural relaxation processes and implemented within an Ansys® environment. The volume relaxation kinetics was modeled by coupling the KAHR (Kovacs, Aklonis, Hutchinson, Ramos) phenomenological theory [4] with the linear viscoelastic theory [5-7]. The numerical algorithm translates the specific volume theoretical predictions at each body point as applied non-mechanical loads acting on the component. The viscoelastic functions were obtained from two simple experimental data, namely the linear viscoelastic response in shear and the PVT (pressure volume temperature) behavior. The dimensionless bulk compliance was extracted from PVT data since it coincides with the memory function appearing in the KAHR phenomenological theory [7]. It is showed that the residual stress scales linearly with the logarithm of the Biot's number.

  2. Distribution of residues and primitive roots

    Indian Academy of Sciences (India)

    Replacing the function f by g, we get the required estimate for N(p, N). D. Proof of Theorem 1.1. When p = 7, we clearly see that (1, 2) is a consecutive pair of quadratic residue modulo 7. Assume that p ≥ 11. If 10 is a quadratic residue modulo p, then we have (9, 10) as a consecutive pair of quadratic residues modulo p, ...

  3. Neutron residual stress measurements in linepipe

    International Nuclear Information System (INIS)

    Law, Michael; Gnaepel-Herold, Thomas; Luzin, Vladimir; Bowie, Graham

    2006-01-01

    Residual stresses in gas pipelines are generated by manufacturing and construction processes and may affect the subsequent pipe integrity. In the present work, the residual stresses in eight samples of linepipe were measured by neutron diffraction. Residual stresses changed with some coating processes. This has special implications in understanding and mitigating stress corrosion cracking, a major safety and economic problem in some gas pipelines

  4. Using cotton plant residue to produce briquettes

    Energy Technology Data Exchange (ETDEWEB)

    Coates, W. [University of Arizona, Tucson, AZ (United States). Bioresources Research Facility

    2000-07-01

    In Arizona, cotton (Gossypium) plant residue left in the field following harvest must be buried to prevent it from serving as an overwintering site for insects such as the pink bollworm. Most tillage operations employed to incorporate the residue into the soil are energy intensive and often degrade soil structure. Trials showed that cotton plant residue could be incorporated with pecan shells to produce commercially acceptable briquettes. Pecan shell briquettes containing cotton residue rather than waste paper were slightly less durable, when made using equivalent weight mixtures and moisture contents. Proximate and ultimate analyses showed the only difference among briquette samples to be a higher ash content in those made using cotton plant residue. Briquettes made with paper demonstrated longer flame out time, and lower ash percentage, compared to those made with cotton plant residue. (author)

  5. Landfilling of waste incineration residues

    DEFF Research Database (Denmark)

    Christensen, Thomas Højlund; Astrup, Thomas; Cai, Zuansi

    2002-01-01

    Residues from waste incineration are bottom ashes and air-pollution-control (APC) residues including fly ashes. The leaching of heavy metals and salts from the ashes is substantial and a wide spectrum of leaching tests and corresponding criteria have been introduced to regulate the landfilling...

  6. Prediction of residue-residue contact matrix for protein-protein interaction with Fisher score features and deep learning.

    Science.gov (United States)

    Du, Tianchuan; Liao, Li; Wu, Cathy H; Sun, Bilin

    2016-11-01

    Protein-protein interactions play essential roles in many biological processes. Acquiring knowledge of the residue-residue contact information of two interacting proteins is not only helpful in annotating functions for proteins, but also critical for structure-based drug design. The prediction of the protein residue-residue contact matrix of the interfacial regions is challenging. In this work, we introduced deep learning techniques (specifically, stacked autoencoders) to build deep neural network models to tackled the residue-residue contact prediction problem. In tandem with interaction profile Hidden Markov Models, which was used first to extract Fisher score features from protein sequences, stacked autoencoders were deployed to extract and learn hidden abstract features. The deep learning model showed significant improvement over the traditional machine learning model, Support Vector Machines (SVM), with the overall accuracy increased by 15% from 65.40% to 80.82%. We showed that the stacked autoencoders could extract novel features, which can be utilized by deep neural networks and other classifiers to enhance learning, out of the Fisher score features. It is further shown that deep neural networks have significant advantages over SVM in making use of the newly extracted features. Copyright © 2016. Published by Elsevier Inc.

  7. Fermion cluster algorithms

    International Nuclear Information System (INIS)

    Chandrasekharan, Shailesh

    2000-01-01

    Cluster algorithms have been recently used to eliminate sign problems that plague Monte-Carlo methods in a variety of systems. In particular such algorithms can also be used to solve sign problems associated with the permutation of fermion world lines. This solution leads to the possibility of designing fermion cluster algorithms in certain cases. Using the example of free non-relativistic fermions we discuss the ideas underlying the algorithm

  8. Recipe for residual oil saturation determination

    Energy Technology Data Exchange (ETDEWEB)

    Guillory, A.J.; Kidwell, C.M.

    1979-01-01

    In 1978, Shell Oil Co., in conjunction with the US Department of Energy, conducted a residual oil saturation study in a deep, hot high-pressured Gulf Coast Reservoir. The work was conducted prior to initiation of CO/sub 2/ tertiary recovery pilot. Many problems had to be resolved prior to and during the residual oil saturation determination. The problems confronted are outlined such that the procedure can be used much like a cookbook in designing future studies in similar reservoirs. Primary discussion centers around planning and results of a log-inject-log operation used as a prime method to determine the residual oil saturation. Several independent methods were used to calculate the residual oil saturation in the subject well in an interval between 12,910 ft (3935 m) and 12,020 ft (3938 m). In general, these numbers were in good agreement and indicated a residual oil saturation between 22% and 24%. 10 references.

  9. 40 CFR 180.564 - Indoxacarb; tolerances for residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Indoxacarb; tolerances for residues...) PESTICIDE PROGRAMS TOLERANCES AND EXEMPTIONS FOR PESTICIDE CHEMICAL RESIDUES IN FOOD Specific Tolerances § 180.564 Indoxacarb; tolerances for residues. (a) General. Tolerances are established for residues of...

  10. Algorithm improvement program nuclide identification algorithm scoring criteria and scoring application.

    Energy Technology Data Exchange (ETDEWEB)

    Enghauser, Michael [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2016-02-01

    The goal of the Domestic Nuclear Detection Office (DNDO) Algorithm Improvement Program (AIP) is to facilitate gamma-radiation detector nuclide identification algorithm development, improvement, and validation. Accordingly, scoring criteria have been developed to objectively assess the performance of nuclide identification algorithms. In addition, a Microsoft Excel spreadsheet application for automated nuclide identification scoring has been developed. This report provides an overview of the equations, nuclide weighting factors, nuclide equivalencies, and configuration weighting factors used by the application for scoring nuclide identification algorithm performance. Furthermore, this report presents a general overview of the nuclide identification algorithm scoring application including illustrative examples.

  11. Residues and duality for projective algebraic varieties

    CERN Document Server

    Kunz, Ernst; Dickenstein, Alicia

    2008-01-01

    This book, which grew out of lectures by E. Kunz for students with a background in algebra and algebraic geometry, develops local and global duality theory in the special case of (possibly singular) algebraic varieties over algebraically closed base fields. It describes duality and residue theorems in terms of K�hler differential forms and their residues. The properties of residues are introduced via local cohomology. Special emphasis is given to the relation between residues to classical results of algebraic geometry and their generalizations. The contribution by A. Dickenstein gives applications of residues and duality to polynomial solutions of constant coefficient partial differential equations and to problems in interpolation and ideal membership. D. A. Cox explains toric residues and relates them to the earlier text. The book is intended as an introduction to more advanced treatments and further applications of the subject, to which numerous bibliographical hints are given.

  12. Recovery of transuranics from process residues

    International Nuclear Information System (INIS)

    Gray, J.H.; Gray, L.W.

    1987-01-01

    Process residues are generated at both the Rocky Flats Plant (RFP) and the Savannah River Plant (SRP) during aqueous chemical and pyrochemical operations. Frequently, process operations will result in either impure products or produce residues sufficiently contaminated with transuranics to be nondiscardable as waste. Purification and recovery flowsheets for process residues have been developed to generate solutions compatible with subsequent Purex operations and either solid or liquid waste suitable for disposal. The ''scrub alloy'' and the ''anode heel alloy'' are examples of materials generated at RFP which have been processed at SRP using the developed recovery flowsheets. Examples of process residues being generated at SRP for which flowsheets are under development include LECO crucibles and alpha-contaminated hydraulic oil

  13. Automated backbone assignment of labeled proteins using the threshold accepting algorithm

    International Nuclear Information System (INIS)

    Leutner, Michael; Gschwind, Ruth M.; Liermann, Jens; Schwarz, Christian; Gemmecker, Gerd; Kessler, Horst

    1998-01-01

    The sequential assignment of backbone resonances is the first step in the structure determination of proteins by heteronuclear NMR. For larger proteins, an assignment strategy based on proton side-chain information is no longer suitable for the use in an automated procedure. Our program PASTA (Protein ASsignment by Threshold Accepting) is therefore designed to partially or fully automate the sequential assignment of proteins, based on the analysis of NMR backbone resonances plus C β information. In order to overcome the problems caused by peak overlap and missing signals in an automated assignment process, PASTA uses threshold accepting, a combinatorial optimization strategy, which is superior to simulated annealing due to generally faster convergence and better solutions. The reliability of this algorithm is shown by reproducing the complete sequential backbone assignment of several proteins from published NMR data. The robustness of the algorithm against misassigned signals, noise, spectral overlap and missing peaks is shown by repeating the assignment with reduced sequential information and increased chemical shift tolerances. The performance of the program on real data is finally demonstrated with automatically picked peak lists of human nonpancreatic synovial phospholipase A 2 , a protein with 124 residues

  14. 40 CFR 279.59 - Management of residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 26 2010-07-01 2010-07-01 false Management of residues. 279.59 Section 279.59 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID WASTES (CONTINUED... Management of residues. Owners and operators who generate residues from the storage, processing, or re...

  15. Algorithm of Particle Data Association for SLAM Based on Improved Ant Algorithm

    Directory of Open Access Journals (Sweden)

    KeKe Gen

    2015-01-01

    Full Text Available The article considers a problem of data association algorithm for simultaneous localization and mapping guidelines in determining the route of unmanned aerial vehicles (UAVs. Currently, these equipments are already widely used, but mainly controlled from the remote operator. An urgent task is to develop a control system that allows for autonomous flight. Algorithm SLAM (simultaneous localization and mapping, which allows to predict the location, speed, the ratio of flight parameters and the coordinates of landmarks and obstacles in an unknown environment, is one of the key technologies to achieve real autonomous UAV flight. The aim of this work is to study the possibility of solving this problem by using an improved ant algorithm.The data association for SLAM algorithm is meant to establish a matching set of observed landmarks and landmarks in the state vector. Ant algorithm is one of the widely used optimization algorithms with positive feedback and the ability to search in parallel, so the algorithm is suitable for solving the problem of data association for SLAM. But the traditional ant algorithm in the process of finding routes easily falls into local optimum. Adding random perturbations in the process of updating the global pheromone to avoid local optima. Setting limits pheromone on the route can increase the search space with a reasonable amount of calculations for finding the optimal route.The paper proposes an algorithm of the local data association for SLAM algorithm based on an improved ant algorithm. To increase the speed of calculation, local data association is used instead of the global data association. The first stage of the algorithm defines targets in the matching space and the observed landmarks with the possibility of association by the criterion of individual compatibility (IC. The second stage defines the matched landmarks and their coordinates using improved ant algorithm. Simulation results confirm the efficiency and

  16. Residual stress characterization of steel TIG welds by neutron diffraction and by residual magnetic stray field mappings

    Energy Technology Data Exchange (ETDEWEB)

    Stegemann, Robert, E-mail: Robert.Stegemann@bam.de [Bundesanstalt für Materialforschung und -prüfung (BAM), Unter den Eichen 87, 12200 Berlin (Germany); Cabeza, Sandra; Lyamkin, Viktor; Bruno, Giovanni; Pittner, Andreas [Bundesanstalt für Materialforschung und -prüfung (BAM), Unter den Eichen 87, 12200 Berlin (Germany); Wimpory, Robert; Boin, Mirko [HZB Helmholtz-Zentrum Berlin, Hahn-Meitner-Platz 1, 14109 Berlin (Germany); Kreutzbruck, Marc [Bundesanstalt für Materialforschung und -prüfung (BAM), Unter den Eichen 87, 12200 Berlin (Germany); IKT, University of Stuttgart, Pfaffenwaldring 32, 70569 Stuttgart (Germany)

    2017-03-15

    The residual stress distribution of tungsten inert gas welded S235JRC+C plates was determined by means of neutron diffraction (ND). Large longitudinal residual stresses with maxima around 600 MPa were found. With these results as reference, the evaluation of residual stress with high spatial resolution GMR (giant magneto resistance) sensors was discussed. The experiments performed indicate a correlation between changes in residual stresses (ND) and the normal component of local residual magnetic stray fields (GMR). Spatial variations in the magnetic field strength perpendicular to the welds are in the order of the magnetic field of the earth. - Highlights: • Comparison of magnetic microstructure with neutron diffraction stress analysis. • High spatial resolution magnetic stray field images of hypereutectoid TIG welds. • Spatial variations of the stray fields are below the magnetic field of the earth. • GMR spin valve gradiometer arrays adapted for the evaluation of magnetic microstructures. • Magnetic stray fields are closely linked to microstructure of the material.

  17. The BR eigenvalue algorithm

    Energy Technology Data Exchange (ETDEWEB)

    Geist, G.A. [Oak Ridge National Lab., TN (United States). Computer Science and Mathematics Div.; Howell, G.W. [Florida Inst. of Tech., Melbourne, FL (United States). Dept. of Applied Mathematics; Watkins, D.S. [Washington State Univ., Pullman, WA (United States). Dept. of Pure and Applied Mathematics

    1997-11-01

    The BR algorithm, a new method for calculating the eigenvalues of an upper Hessenberg matrix, is introduced. It is a bulge-chasing algorithm like the QR algorithm, but, unlike the QR algorithm, it is well adapted to computing the eigenvalues of the narrowband, nearly tridiagonal matrices generated by the look-ahead Lanczos process. This paper describes the BR algorithm and gives numerical evidence that it works well in conjunction with the Lanczos process. On the biggest problems run so far, the BR algorithm beats the QR algorithm by a factor of 30--60 in computing time and a factor of over 100 in matrix storage space.

  18. RESIDUAL RISK ASSESSMENT: ETHYLENE OXIDE ...

    Science.gov (United States)

    This document describes the residual risk assessment for the Ethylene Oxide Commercial Sterilization source category. For stationary sources, section 112 (f) of the Clean Air Act requires EPA to assess risks to human health and the environment following implementation of technology-based control standards. If these technology-based control standards do not provide an ample margin of safety, then EPA is required to promulgate addtional standards. This document describes the methodology and results of the residual risk assessment performed for the Ethylene Oxide Commercial Sterilization source category. The results of this analyiss will assist EPA in determining whether a residual risk rule for this source category is appropriate.

  19. On the parametrization of the Δ residue function

    International Nuclear Information System (INIS)

    Vasan, S.S.

    1976-01-01

    The complex residues at the Δ(1236) and Δ(1950) poles in the relevant partial-wave amplitudes provide information on the behaviour of the Δ Regge residue function in the resonance region u>0. Attempts to incorporate this information in parametrizations of the residue by functions that are real on the real u-axis result in residues which have unsatisfactory behaviour in the region u 2 . The choice of complex functions for the trajectory and residue removes this undesirable feature and provides a better representation of the residue in the resonance region, suggesting that complex parametrizations would be better suited to Regge analyses of near-backward scattering. (Auth.)

  20. Parametrization of the. delta. residue function. [Complex functions

    Energy Technology Data Exchange (ETDEWEB)

    Vasan, S S [Carnegie-Mellon Univ., Pittsburgh, Pa. (USA). Dept. of Physics

    1976-04-19

    The complex residues at the ..delta..(1236) and ..delta..(1950) poles in the relevant partial-wave amplitudes provide information on the behaviour of the ..delta.. Regge residue function in the resonance region u>0. Attempts to incorporate this information in parametrizations of the residue by functions that are real on the real u-axis result in residues which have unsatisfactory behaviour in the region u<1GeV/sup 2/. The choice of complex functions for the trajectory and residue removes this undesirable feature and provides a better representation of the residue in the resonance region, suggesting that complex parametrizations would be better suited to Regge analyses of near-backward scattering.

  1. Long-term stabilization of crop residues and soil organic carbon affected by residue quality and initial soil pH.

    Science.gov (United States)

    Wang, Xiaojuan; Butterly, Clayton R; Baldock, Jeff A; Tang, Caixian

    2017-06-01

    Residues differing in quality and carbon (C) chemistry are presumed to contribute differently to soil pH change and long-term soil organic carbon (SOC) pools. This study examined the liming effect of different crop residues (canola, chickpea and wheat) down the soil profile (0-30cm) in two sandy soils differing in initial pH as well as the long-term stability of SOC at the amended layer (0-10cm) using mid-infrared (MIR) and solid-state 13 C nuclear magnetic resonance (NMR) spectroscopy. A field column experiment was conducted for 48months. Chickpea- and canola-residue amendments increased soil pH at 0-10cm in the Podzol by up to 0.47 and 0.36units, and in the Cambisol by 0.31 and 0.18units, respectively, at 48months when compared with the non-residue-amended control. The decomposition of crop residues was greatly retarded in the Podzol with lower initial soil pH during the first 9months. The MIR-predicted particulate organic C (POC) acted as the major C sink for residue-derived C in the Podzol. In contrast, depletion of POC and recovery of residue C in MIR-predicted humic organic C (HOC) were detected in the Cambisol within 3months. Residue types showed little impact on total SOC and its chemical composition in the Cambisol at 48months, in contrast to the Podzol. The final HOC and resistant organic C (ROC) pools in the Podzol amended with canola and chickpea residues were about 25% lower than the control. This apparent priming effect might be related to the greater liming effect of these two residues in the Podzol. Copyright © 2017 Elsevier B.V. All rights reserved.

  2. 40 CFR 279.67 - Management of residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 26 2010-07-01 2010-07-01 false Management of residues. 279.67 Section 279.67 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID WASTES (CONTINUED... for Energy Recovery § 279.67 Management of residues. Burners who generate residues from the storage or...

  3. 40 CFR 279.47 - Management of residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 26 2010-07-01 2010-07-01 false Management of residues. 279.47 Section 279.47 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID WASTES (CONTINUED... Management of residues. Transporters who generate residues from the storage or transport of used oil must...

  4. OGT (O-GlcNAc Transferase) Selectively Modifies Multiple Residues Unique to Lamin A.

    Science.gov (United States)

    Simon, Dan N; Wriston, Amanda; Fan, Qiong; Shabanowitz, Jeffrey; Florwick, Alyssa; Dharmaraj, Tejas; Peterson, Sherket B; Gruenbaum, Yosef; Carlson, Cathrine R; Grønning-Wang, Line M; Hunt, Donald F; Wilson, Katherine L

    2018-05-17

    The LMNA gene encodes lamins A and C with key roles in nuclear structure, signaling, gene regulation, and genome integrity. Mutations in LMNA cause over 12 diseases ('laminopathies'). Lamins A and C are identical for their first 566 residues. However, they form separate filaments in vivo, with apparently distinct roles. We report that lamin A is β- O -linked N -acetylglucosamine- (O -GlcNAc)-modified in human hepatoma (Huh7) cells and in mouse liver. In vitro assays with purified O -GlcNAc transferase (OGT) enzyme showed robust O -GlcNAcylation of recombinant mature lamin A tails (residues 385⁻646), with no detectable modification of lamin B1, lamin C, or 'progerin' (Δ50) tails. Using mass spectrometry, we identified 11 O -GlcNAc sites in a 'sweet spot' unique to lamin A, with up to seven sugars per peptide. Most sites were unpredicted by current algorithms. Double-mutant (S612A/T643A) lamin A tails were still robustly O -GlcNAc-modified at seven sites. By contrast, O -GlcNAcylation was undetectable on tails bearing deletion Δ50, which causes Hutchinson⁻Gilford progeria syndrome, and greatly reduced by deletion Δ35. We conclude that residues deleted in progeria are required for substrate recognition and/or modification by OGT in vitro. Interestingly, deletion Δ35, which does not remove the majority of identified O -GlcNAc sites, does remove potential OGT-association motifs (lamin A residues 622⁻625 and 639⁻645) homologous to that in mouse Tet1. These biochemical results are significant because they identify a novel molecular pathway that may profoundly influence lamin A function. The hypothesis that lamin A is selectively regulated by OGT warrants future testing in vivo, along with two predictions: genetic variants may contribute to disease by perturbing OGT-dependent regulation, and nutrient or other stresses might cause OGT to misregulate wildtype lamin A.

  5. Fluorescence imaging to quantify crop residue cover

    Science.gov (United States)

    Daughtry, C. S. T.; Mcmurtrey, J. E., III; Chappelle, E. W.

    1994-01-01

    Crop residues, the portion of the crop left in the field after harvest, can be an important management factor in controlling soil erosion. Methods to quantify residue cover are needed that are rapid, accurate, and objective. Scenes with known amounts of crop residue were illuminated with long wave ultraviolet (UV) radiation and fluorescence images were recorded with an intensified video camera fitted with a 453 to 488 nm band pass filter. A light colored soil and a dark colored soil were used as background for the weathered soybean stems. Residue cover was determined by counting the proportion of the pixels in the image with fluorescence values greater than a threshold. Soil pixels had the lowest gray levels in the images. The values of the soybean residue pixels spanned nearly the full range of the 8-bit video data. Classification accuracies typically were within 3(absolute units) of measured cover values. Video imaging can provide an intuitive understanding of the fraction of the soil covered by residue.

  6. Geometric approximation algorithms

    CERN Document Server

    Har-Peled, Sariel

    2011-01-01

    Exact algorithms for dealing with geometric objects are complicated, hard to implement in practice, and slow. Over the last 20 years a theory of geometric approximation algorithms has emerged. These algorithms tend to be simple, fast, and more robust than their exact counterparts. This book is the first to cover geometric approximation algorithms in detail. In addition, more traditional computational geometry techniques that are widely used in developing such algorithms, like sampling, linear programming, etc., are also surveyed. Other topics covered include approximate nearest-neighbor search, shape approximation, coresets, dimension reduction, and embeddings. The topics covered are relatively independent and are supplemented by exercises. Close to 200 color figures are included in the text to illustrate proofs and ideas.

  7. Vesícula residual

    Directory of Open Access Journals (Sweden)

    Júlio C. U. Coelho

    Full Text Available Our objective is to report three patients with recurrent severe upper abdominal pain secondary to residual gallbladder. All patients had been subjected to cholecystectomy from 1 to 20 years before. The diagnosis was established after several episodes of severe upper abdominal pain by imaging exams: ultrasonography, tomography, or endoscopic retrograde cholangiography. Removal of the residual gallbladder led to complete resolution of symptoms. Partial removal of the gallbladder is a very rare cause of postcholecystectomy symptoms.

  8. Polychlorinated Biphenyls (PCB) Residue Effects Database

    Data.gov (United States)

    U.S. Environmental Protection Agency — The PCB Residue Effects (PCBRes) Database was developed to assist scientists and risk assessors in correlating PCB and dioxin-like compound residues with toxic...

  9. Pesticide residues in locally available cereals and vegetables

    International Nuclear Information System (INIS)

    Cunanan, S.A.; Santos, F.L.; Bonoan, L.S.

    1976-03-01

    Vegetable samples (pechay, cabbage, lettuce, green beans and tomatoes) bought from public markets in the Metro Manila area were analyzed for pesticide residues using gas chromatography. The samples analyzed in 1968-69 contained high levels of chlorinated pesticides such as DDT, Aldrin, Endrin, and Thiodan, while in the samples analyzed in January 1976, no chlorinated and organophosphate pesticides were detected. Cereal samples (rice, corn and sorghum) were obtained from the National Grains Authority and analyzed for pesticide residues and bromine residues. Total bromine residues was determined by neutron activation analysis. In most of the samples analyzed, the concentrations of pesticide residues were below the tolerance levels set by the FAO/WHO Committee on Pesticide Residues in Foods. An exception was one rice sample from Thailand, the bromine residue content (110ppm) of which exceeds the tolerance level of 50ppm

  10. 40 CFR 721.9635 - Terpene residue distillates.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 30 2010-07-01 2010-07-01 false Terpene residue distillates. 721.9635... Substances § 721.9635 Terpene residue distillates. (a) Chemical substance and significant new uses subject to reporting. (1) The chemical substance identified generically as terpene residue distillates (PMN P-96-897...

  11. Efficient particle filtering through residual nudging

    KAUST Repository

    Luo, Xiaodong

    2013-05-15

    We introduce an auxiliary technique, called residual nudging, to the particle filter to enhance its performance in cases where it performs poorly. The main idea of residual nudging is to monitor and, if necessary, adjust the residual norm of a state estimate in the observation space so that it does not exceed a pre-specified threshold. We suggest a rule to choose the pre-specified threshold, and construct a state estimate accordingly to achieve this objective. Numerical experiments suggest that introducing residual nudging to a particle filter may (substantially) improve its performance, in terms of filter accuracy and/or stability against divergence, especially when the particle filter is implemented with a relatively small number of particles. © 2013 Royal Meteorological Society.

  12. Algorithms in Singular

    Directory of Open Access Journals (Sweden)

    Hans Schonemann

    1996-12-01

    Full Text Available Some algorithms for singularity theory and algebraic geometry The use of Grobner basis computations for treating systems of polynomial equations has become an important tool in many areas. This paper introduces of the concept of standard bases (a generalization of Grobner bases and the application to some problems from algebraic geometry. The examples are presented as SINGULAR commands. A general introduction to Grobner bases can be found in the textbook [CLO], an introduction to syzygies in [E] and [St1]. SINGULAR is a computer algebra system for computing information about singularities, for use in algebraic geometry. The basic algorithms in SINGULAR are several variants of a general standard basis algorithm for general monomial orderings (see [GG]. This includes wellorderings (Buchberger algorithm ([B1], [B2] and tangent cone orderings (Mora algorithm ([M1], [MPT] as special cases: It is able to work with non-homogeneous and homogeneous input and also to compute in the localization of the polynomial ring in 0. Recent versions include algorithms to factorize polynomials and a factorizing Grobner basis algorithm. For a complete description of SINGULAR see [Si].

  13. Multimodal optimization by using hybrid of artificial bee colony algorithm and BFGS algorithm

    Science.gov (United States)

    Anam, S.

    2017-10-01

    Optimization has become one of the important fields in Mathematics. Many problems in engineering and science can be formulated into optimization problems. They maybe have many local optima. The optimization problem with many local optima, known as multimodal optimization problem, is how to find the global solution. Several metaheuristic methods have been proposed to solve multimodal optimization problems such as Particle Swarm Optimization (PSO), Genetics Algorithm (GA), Artificial Bee Colony (ABC) algorithm, etc. The performance of the ABC algorithm is better than or similar to those of other population-based algorithms with the advantage of employing a fewer control parameters. The ABC algorithm also has the advantages of strong robustness, fast convergence and high flexibility. However, it has the disadvantages premature convergence in the later search period. The accuracy of the optimal value cannot meet the requirements sometimes. Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm is a good iterative method for finding a local optimum. Compared with other local optimization methods, the BFGS algorithm is better. Based on the advantages of the ABC algorithm and the BFGS algorithm, this paper proposes a hybrid of the artificial bee colony algorithm and the BFGS algorithm to solve the multimodal optimization problem. The first step is that the ABC algorithm is run to find a point. In the second step is that the point obtained by the first step is used as an initial point of BFGS algorithm. The results show that the hybrid method can overcome from the basic ABC algorithm problems for almost all test function. However, if the shape of function is flat, the proposed method cannot work well.

  14. Use of ultrasound in petroleum residue upgradation

    Energy Technology Data Exchange (ETDEWEB)

    Sawarkar, A.N.; Pandit, A.B.; Samant, S.D.; Joshi, J.B. [Mumbai Univ., Mumbai (India). Inst. of Chemical Technology

    2009-06-15

    The importance of bottom-of-the barrel upgrading has increased in the current petroleum refining scenario because of the progressively heavier nature of crude oil. Heavy residues contain large concentrations of metals such as vanadium and nickel which foul catalysts and reduce the potential effect of residue fluidized catalytic cracking. This study showed that the cavitational energy induced by ultrasound be be successfully used to upgrade hydrocarbon mixtures. Conventional processes for the upgrading of residual feedstocks, such as thermal cracking and catalytic cracking, were carried out in the temperature range of 400-520 degrees C. Experiments were performed on 2 vacuum residues, Arabian mix vacuum residue (AMVR) and Bombay high vacuum residue (BHVR) and 1 Haldia asphalt (HA). These were subjected to acoustic cavitation for different reaction times from 15 to 120 minutes at ambient temperature and pressure. Two acoustic cavitation devices were compared, namely the ultrasonic bath and ultrasonic horn. In particular, this study compared the ability of these 2 devices to upgrade the petroleum residues to lighter, more value-added products. Different surfactants were used to examine the effect of ultrasound on upgrading the residue when emulsified in water. In order to better understand the reaction mechanism, a kinetic model was developed based on the constituents of the residue. The ultrasonic horn was found to be more effective in bringing about the upgrading than ultrasonic bath. The study also showed that the acoustic cavitation of the aqueous emulsified hydrocarbon mixture could reduce the asphaltenes content to a greater extent than the acoustic cavitation of non-emulsified hydrocarbon mixture. 20 refs., 11 tabs., 17 figs.

  15. Western Canadian wood residue production and consumption trends

    International Nuclear Information System (INIS)

    McCloy, B.

    2006-01-01

    This presentation considered various trends in western Canadian wood residue production and consumption. Potential markets for wood residue products were also discussed. Trends were reviewed by province for the years 2000-2004. British Columbia (BC) is currently the largest producer of residue in the country, and also retains the largest surpluses of bark, sawdust and shavings. Wood residues in BC are used in pulp and plywood mill production, as well as in the creation of particleboard and MDF. Surplus mill wood residue production in the province has greatly increased due to the Mountain Pine Beetle (MPB) infestation, which has in turn spurred expansion of the BC interior sawmill industry. The infestation has also resulted in a glut of pulp chips. Current wood residue products in Alberta are mostly used in pulp mill combined heat and power (CHP) systems, as well as for wood pellet production and the creation of particleboard and MDF. It was noted that surplus residues are rapidly declining in the province. Saskatchewan's wood residue storage piles are estimated to contain 2,900,000 BDt, while Manitoba surpluses are relatively minor. It was suggested that high natural gas prices have increased the payback on wood energy systems to approximately 2 years. The value of wood residue is now greater than $100/BDt as a substitute for natural gas once the wood energy system has been fully depreciated. Sawmills may now wish to consider equipping themselves to sell wood residue products, as most sawmills only require 20 per cent of their residues for heating purposes. It was concluded that markets for hog fuel wood pellets should be developed in Canada and internationally. Future markets may also develop if natural gas currently used in pulp mill power boilers and lime kilns is replaced with wood residue energy systems. refs., tabs., figs

  16. Modified Firefly Algorithm

    Directory of Open Access Journals (Sweden)

    Surafel Luleseged Tilahun

    2012-01-01

    Full Text Available Firefly algorithm is one of the new metaheuristic algorithms for optimization problems. The algorithm is inspired by the flashing behavior of fireflies. In the algorithm, randomly generated solutions will be considered as fireflies, and brightness is assigned depending on their performance on the objective function. One of the rules used to construct the algorithm is, a firefly will be attracted to a brighter firefly, and if there is no brighter firefly, it will move randomly. In this paper we modify this random movement of the brighter firefly by generating random directions in order to determine the best direction in which the brightness increases. If such a direction is not generated, it will remain in its current position. Furthermore the assignment of attractiveness is modified in such a way that the effect of the objective function is magnified. From the simulation result it is shown that the modified firefly algorithm performs better than the standard one in finding the best solution with smaller CPU time.

  17. Persistence of pirimiphos methyl residues in stored wheat

    International Nuclear Information System (INIS)

    Zennouche, B.; Bennaceur, M.

    1991-03-01

    The persistence of 14C pirimiphos methyl in stored wheat was investigated under local conditions during a storage period of 6 months. Surface, methanol extractable and bound residues were quantified. There was a gradual increase in the surface residues over the storage period. 27,97% of the applied amount was found in the surface residues and 54,16% in the extractable and bound residues one month after treatment. After 6 months storage, about 17,8% of the total quantity of the applied pirimiphos methyl remained on the surface and about 58,6% was present in the grains (extractable and bound residues). During cooking 10,91% of the extractable and bound residues were lost

  18. Pesticide residues in birds and mammals

    Science.gov (United States)

    Stickel, L.F.; Edwards, C.A.

    1973-01-01

    SUMMARY: Residues of organochlorine pesticides and their breakdown products are present in the tissues of essentially all wild birds throughout the world. These chemicals accumulate in fat from a relatively small environmental exposure. DDE and dieldrin are most prevalent. Others, such as heptachlor epoxide, chlordane, endrin, and benzene hexachloride also occur, the quantities and kinds generally reflecting local or regional use. Accumulation may be sufficient to kill animals following applications for pest control. This has occurred in several large-scale programmes in the United States. Mortality has also resulted from unintentional leakage of chemical from commercial establishments. Residues may persist in the environment for many years, exposing successive generations of animals. In general, birds that eat other birds, or fish, have higher residues than those that eat seeds and vegetation. The kinetic processes of absorption, metabolism, storage, and output differ according to both kind of chemical and species of animal. When exposure is low and continuous, a balance between intake and excretion may be achieved. Residues reach a balance at an approximate animal body equilibrium or plateau; the storage is generally proportional to dose. Experiments with chickens show that dieldrin and heptachlor epoxide have the greatest propensity for storage, endrin next, then DDT, then lindane. The storage of DDT was complicated by its metabolism to DDE and DDD, but other studies show that DDE has a much greater propensity for storage than either DDD or DDT. Methoxychlor has little cumulative capacity in birds. Residues in eggs reflect and parallel those in the parent bird during accumulation, equilibrium, and decline when dosage is discontinued. Residues with the greatest propensity for storage are also lost most slowly. Rate of loss of residues can be modified by dietary components and is speeded by weight loss of the animal. Under sublethal conditions of continuous

  19. Carbaryl residues in maize and processed products

    International Nuclear Information System (INIS)

    Qureshi, M.J.; Sattar, A. Jr.; Naqvi, M.H.

    1981-01-01

    Carbaryl residues in two local maize varieties were determined using a colorimetric method. No significant differences were observed for residues of the two varieties which ranged between 12.0 to 13.75 mg/kg in the crude oil, and averaged 1.04 and 0.67 mg/kg in the flour and cake respectively. In whole maize plants, carbaryl residues declined to approximately 2 mg/kg 35 days after treatment. Cooking in aqueous, oil or aqueous-oil media led to 63-83% loss of carbaryl residues, after 30 minutes. (author)

  20. Network-Oblivious Algorithms

    DEFF Research Database (Denmark)

    Bilardi, Gianfranco; Pietracaprina, Andrea; Pucci, Geppino

    2016-01-01

    A framework is proposed for the design and analysis of network-oblivious algorithms, namely algorithms that can run unchanged, yet efficiently, on a variety of machines characterized by different degrees of parallelism and communication capabilities. The framework prescribes that a network......-oblivious algorithm be specified on a parallel model of computation where the only parameter is the problem’s input size, and then evaluated on a model with two parameters, capturing parallelism granularity and communication latency. It is shown that for a wide class of network-oblivious algorithms, optimality...... of cache hierarchies, to the realm of parallel computation. Its effectiveness is illustrated by providing optimal network-oblivious algorithms for a number of key problems. Some limitations of the oblivious approach are also discussed....

  1. A Discrete-Time Algorithm for Stiffness Extraction from sEMG and Its Application in Antidisturbance Teleoperation

    Directory of Open Access Journals (Sweden)

    Peidong Liang

    2016-01-01

    Full Text Available We have developed a new discrete-time algorithm of stiffness extraction from muscle surface electromyography (sEMG collected from human operator’s arms and have applied it for antidisturbance control in robot teleoperation. The variation of arm stiffness is estimated from sEMG signals and transferred to a telerobot under variable impedance control to imitate human motor control behaviours, particularly for disturbance attenuation. In comparison to the estimation of stiffness from sEMG, the proposed algorithm is able to reduce the nonlinear residual error effect and to enhance robustness and to simplify stiffness calibration. In order to extract a smoothing stiffness enveloping from sEMG signals, two enveloping methods are employed in this paper, namely, fast linear enveloping based on low pass filtering and moving average and amplitude monocomponent and frequency modulating (AM-FM method. Both methods have been incorporated into the proposed stiffness variance estimation algorithm and are extensively tested. The test results show that stiffness variation extraction based on the two methods is sensitive and robust to attenuation disturbance. It could potentially be applied for teleoperation in the presence of hazardous surroundings or human robot physical cooperation scenarios.

  2. A novel hybrid algorithm of GSA with Kepler algorithm for numerical optimization

    Directory of Open Access Journals (Sweden)

    Soroor Sarafrazi

    2015-07-01

    Full Text Available It is now well recognized that pure algorithms can be promisingly improved by hybridization with other techniques. One of the relatively new metaheuristic algorithms is Gravitational Search Algorithm (GSA which is based on the Newton laws. In this paper, to enhance the performance of GSA, a novel algorithm called “Kepler”, inspired by the astrophysics, is introduced. The Kepler algorithm is based on the principle of the first Kepler law. The hybridization of GSA and Kepler algorithm is an efficient approach to provide much stronger specialization in intensification and/or diversification. The performance of GSA–Kepler is evaluated by applying it to 14 benchmark functions with 20–1000 dimensions and the optimal approximation of linear system as a practical optimization problem. The results obtained reveal that the proposed hybrid algorithm is robust enough to optimize the benchmark functions and practical optimization problems.

  3. Prediction of residual stress distributions due to surface machining and welding and crack growth simulation under residual stress distribution

    International Nuclear Information System (INIS)

    Ihara, Ryohei; Katsuyama, JInya; Onizawa, Kunio; Hashimoto, Tadafumi; Mikami, Yoshiki; Mochizuki, Masahito

    2011-01-01

    Research highlights: → Residual stress distributions due to welding and machining are evaluated by XRD and FEM. → Residual stress due to machining shows higher tensile stress than welding near the surface. → Crack growth analysis is performed using calculated residual stress. → Crack growth result is affected machining rather than welding. → Machining is an important factor for crack growth. - Abstract: In nuclear power plants, stress corrosion cracking (SCC) has been observed near the weld zone of the core shroud and primary loop recirculation (PLR) pipes made of low-carbon austenitic stainless steel Type 316L. The joining process of pipes usually includes surface machining and welding. Both processes induce residual stresses, and residual stresses are thus important factors in the occurrence and propagation of SCC. In this study, the finite element method (FEM) was used to estimate residual stress distributions generated by butt welding and surface machining. The thermoelastic-plastic analysis was performed for the welding simulation, and the thermo-mechanical coupled analysis based on the Johnson-Cook material model was performed for the surface machining simulation. In addition, a crack growth analysis based on the stress intensity factor (SIF) calculation was performed using the calculated residual stress distributions that are generated by welding and surface machining. The surface machining analysis showed that tensile residual stress due to surface machining only exists approximately 0.2 mm from the machined surface, and the surface residual stress increases with cutting speed. The crack growth analysis showed that the crack depth is affected by both surface machining and welding, and the crack length is more affected by surface machining than by welding.

  4. Algorithmically specialized parallel computers

    CERN Document Server

    Snyder, Lawrence; Gannon, Dennis B

    1985-01-01

    Algorithmically Specialized Parallel Computers focuses on the concept and characteristics of an algorithmically specialized computer.This book discusses the algorithmically specialized computers, algorithmic specialization using VLSI, and innovative architectures. The architectures and algorithms for digital signal, speech, and image processing and specialized architectures for numerical computations are also elaborated. Other topics include the model for analyzing generalized inter-processor, pipelined architecture for search tree maintenance, and specialized computer organization for raster

  5. Optimization of Proton CT Detector System and Image Reconstruction Algorithm for On-Line Proton Therapy.

    Directory of Open Access Journals (Sweden)

    Chae Young Lee

    Full Text Available The purposes of this study were to optimize a proton computed tomography system (pCT for proton range verification and to confirm the pCT image reconstruction algorithm based on projection images generated with optimized parameters. For this purpose, we developed a new pCT scanner using the Geometry and Tracking (GEANT 4.9.6 simulation toolkit. GEANT4 simulations were performed to optimize the geometric parameters representing the detector thickness and the distance between the detectors for pCT. The system consisted of four silicon strip detectors for particle tracking and a calorimeter to measure the residual energies of the individual protons. The optimized pCT system design was then adjusted to ensure that the solution to a CS-based convex optimization problem would converge to yield the desired pCT images after a reasonable number of iterative corrections. In particular, we used a total variation-based formulation that has been useful in exploiting prior knowledge about the minimal variations of proton attenuation characteristics in the human body. Examinations performed using our CS algorithm showed that high-quality pCT images could be reconstructed using sets of 72 projections within 20 iterations and without any streaks or noise, which can be caused by under-sampling and proton starvation. Moreover, the images yielded by this CS algorithm were found to be of higher quality than those obtained using other reconstruction algorithms. The optimized pCT scanner system demonstrated the potential to perform high-quality pCT during on-line image-guided proton therapy, without increasing the imaging dose, by applying our CS based proton CT reconstruction algorithm. Further, we make our optimized detector system and CS-based proton CT reconstruction algorithm potentially useful in on-line proton therapy.

  6. Residual analysis for spatial point processes

    DEFF Research Database (Denmark)

    Baddeley, A.; Turner, R.; Møller, Jesper

    We define residuals for point process models fitted to spatial point pattern data, and propose diagnostic plots based on these residuals. The techniques apply to any Gibbs point process model, which may exhibit spatial heterogeneity, interpoint interaction and dependence on spatial covariates. Ou...... or covariate effects. Q-Q plots of the residuals are effective in diagnosing interpoint interaction. Some existing ad hoc statistics of point patterns (quadrat counts, scan statistic, kernel smoothed intensity, Berman's diagnostic) are recovered as special cases....

  7. Formation of failure matrix and failure–free control algorithm for multi–sectioned Switched–reluctance drive

    International Nuclear Information System (INIS)

    Odnokopylov, G; Rozayev, I

    2014-01-01

    We review fault-tolerant switched reluctance drive with sectioning of the three–phase stator winding. In the operating process of an electric drive, there will be continuous monitoring of the operating state on the basis of a developed algorithm to analyse drive operability and formation tabulate a failure matrix. The paper introduces a failure–free control algorithm for multi–section switch – reluctance motor with formation the assignment values of amplitude phase currents taking into account the failure matrix. We show that in an emergency such single failure or multiple failure in switched–reluctance drive it is possible to provide reduction of torque fall and pro–gressively stock depletion with providing fault–tolerance of drive system. A method of residual life evaluation is proposed on the basis of calculating the coefficient of operability of the electric drive system that gives possibility to control the endurance of electric drive in real time from operational to completely unusable

  8. DEVELOPMENT OF A NEW ALGORITHM FOR KEY AND S-BOX GENERATION IN BLOWFISH ALGORITHM

    Directory of Open Access Journals (Sweden)

    TAYSEER S. ATIA

    2014-08-01

    Full Text Available Blowfish algorithm is a block cipher algorithm, its strong, simple algorithm used to encrypt data in block of size 64-bit. Key and S-box generation process in this algorithm require time and memory space the reasons that make this algorithm not convenient to be used in smart card or application requires changing secret key frequently. In this paper a new key and S-box generation process was developed based on Self Synchronization Stream Cipher (SSS algorithm where the key generation process for this algorithm was modified to be used with the blowfish algorithm. Test result shows that the generation process requires relatively slow time and reasonably low memory requirement and this enhance the algorithm and gave it the possibility for different usage.

  9. Monte Carlo algorithms with absorbing Markov chains: Fast local algorithms for slow dynamics

    International Nuclear Information System (INIS)

    Novotny, M.A.

    1995-01-01

    A class of Monte Carlo algorithms which incorporate absorbing Markov chains is presented. In a particular limit, the lowest order of these algorithms reduces to the n-fold way algorithm. These algorithms are applied to study the escape from the metastable state in the two-dimensional square-lattice nearest-neighbor Ising ferromagnet in an unfavorable applied field, and the agreement with theoretical predictions is very good. It is demonstrated that the higher-order algorithms can be many orders of magnitude faster than either the traditional Monte Carlo or n-fold way algorithms

  10. Evaluation of heavy metals in hazardous automobile shredder residue thermal residue and immobilization with novel nano-size calcium dispersed reagent.

    Science.gov (United States)

    Lee, Chi-Hyeon; Truc, Nguyen Thi Thanh; Lee, Byeong-Kyu; Mitoma, Yoshiharu; Mallampati, Srinivasa Reddy

    2015-10-15

    This study was conducted to synthesize and apply a nano-size calcium dispersed reagent as an immobilization material for heavy metal-contaminated automobile shredder residues (ASR) dust/thermal residues in dry condition. Simple mixing with a nanometallic Ca/CaO/PO4 dispersion mixture immobilized 95-100% of heavy metals in ASR dust/thermal residues (including bottom ash, cavity ash, boiler and bag filter ash). The quantity of heavy metals leached from thermal residues after treatment by nanometallic Ca/CaO/PO4 was lower than the Korean standard regulatory limit for hazardous waste landfills. The morphology and elemental composition of the nanometallic Ca/CaO-treated ASR residue were characterized by field emission scanning election microscopy combined with electron dispersive spectroscopy (FE-SEM/EDS). The results indicated that the amounts of heavy metals detectable on the ASR thermal residue surface decreased and the Ca/PO4 mass percent increased. X-ray diffraction (XRD) pattern analysis indicated that the main fraction of enclosed/bound materials on ASR residue included Ca/PO4- associated crystalline complexes, and that immobile Ca/PO4 salts remarkably inhibited the desorption of heavy metals from ASR residues. These results support the potential use of nanometallic Ca/CaO/PO4 as a simple, suitable and highly efficient material for the gentle immobilization of heavy metals in hazardous ASR thermal residue in dry condition. Copyright © 2015 Elsevier B.V. All rights reserved.

  11. Dynamic route guidance algorithm based algorithm based on artificial immune system

    Institute of Scientific and Technical Information of China (English)

    2007-01-01

    To improve the performance of the K-shortest paths search in intelligent traffic guidance systems,this paper proposes an optimal search algorithm based on the intelligent optimization search theory and the memphor mechanism of vertebrate immune systems.This algorithm,applied to the urban traffic network model established by the node-expanding method,can expediently realize K-shortest paths search in the urban traffic guidance systems.Because of the immune memory and global parallel search ability from artificial immune systems,K shortest paths can be found without any repeat,which indicates evidently the superiority of the algorithm to the conventional ones.Not only does it perform a better parallelism,the algorithm also prevents premature phenomenon that often occurs in genetic algorithms.Thus,it is especially suitable for real-time requirement of the traffic guidance system and other engineering optimal applications.A case study verifies the efficiency and the practicability of the algorithm aforementioned.

  12. Hamiltonian Algorithm Sound Synthesis

    OpenAIRE

    大矢, 健一

    2013-01-01

    Hamiltonian Algorithm (HA) is an algorithm for searching solutions is optimization problems. This paper introduces a sound synthesis technique using Hamiltonian Algorithm and shows a simple example. "Hamiltonian Algorithm Sound Synthesis" uses phase transition effect in HA. Because of this transition effect, totally new waveforms are produced.

  13. Acceleration of step and linear discontinuous schemes for the method of characteristics in DRAGON5

    Directory of Open Access Journals (Sweden)

    Alain Hébert

    2017-09-01

    Full Text Available The applicability of the algebraic collapsing acceleration (ACA technique to the method of characteristics (MOC in cases with scattering anisotropy and/or linear sources was investigated. Previously, the ACA was proven successful in cases with isotropic scattering and uniform (step sources. A presentation is first made of the MOC implementation, available in the DRAGON5 code. Two categories of schemes are available for integrating the propagation equations: (1 the first category is based on exact integration and leads to the classical step characteristics (SC and linear discontinuous characteristics (LDC schemes and (2 the second category leads to diamond differencing schemes of various orders in space. The acceleration of these MOC schemes using a combination of the generalized minimal residual [GMRES(m] method preconditioned with the ACA technique was focused on. Numerical results are provided for a two-dimensional (2D eight-symmetry pressurized water reactor (PWR assembly mockup in the context of the DRAGON5 code.

  14. Influence of organic waste and residue mud additions on chemical, physical and microbial properties of bauxite residue sand.

    Science.gov (United States)

    Jones, Benjamin E H; Haynes, Richard J; Phillips, Ian R

    2011-02-01

    In an alumina refinery, bauxite ore is treated with sodium hydroxide at high temperatures and pressures and for every tone of alumina produced, about 2 tones of alkaline, saline bauxite processing waste is also produced. At Alcoa, a dry stacking system of disposal is used, and it is the sand fraction of the processing waste that is rehabilitated. There is little information available regarding the most appropriate amendments to add to the processing sand to aid in revegetation. The purpose of this study was to investigate how the addition of organic wastes (biosolids and poultry manure), in the presence or absence of added residue mud, would affect the properties of the residue sand and its suitability for revegetation. Samples of freshly deposited residue sand were collected from Alcoa's Kwinana refinery. Samples were treated with phosphogypsum (2% v/v), incubated, and leached. A laboratory experiment was then set up in which the two organic wastes were applied at 0 or the equivalent to 60 tones ha(-1) in combination with residue mud added at rates of 0%, 10% and 20% v/v. Samples were incubated for 8 weeks, after which, key chemical, physical and microbial properties of the residue sand were measured along with seed germination. Additions of residue mud increased exchangeable Na(+), ESP and the pH, and HCO (3) (-) and Na(+) concentrations in saturation paste extracts. Additions of biosolids and poultry manure increased concentrations of extractable P, NH (4) (+) , K, Mg, Cu, Zn, Mn and Fe. Addition of residue mud, in combination with organic wastes, caused a marked decrease in macroporosity and a concomitant increase in mesoporosity, available water holding capacity and the quantity of water held at field capacity. With increasing residue mud additions, the percentage of sample present as sand particles (2 mm diameter) increased; greatest aggregation occurred where a combination of residue mud and poultry manure were added. Stability of aggregates, as measured by

  15. Modified Clipped LMS Algorithm

    Directory of Open Access Journals (Sweden)

    Lotfizad Mojtaba

    2005-01-01

    Full Text Available Abstract A new algorithm is proposed for updating the weights of an adaptive filter. The proposed algorithm is a modification of an existing method, namely, the clipped LMS, and uses a three-level quantization ( scheme that involves the threshold clipping of the input signals in the filter weight update formula. Mathematical analysis shows the convergence of the filter weights to the optimum Wiener filter weights. Also, it can be proved that the proposed modified clipped LMS (MCLMS algorithm has better tracking than the LMS algorithm. In addition, this algorithm has reduced computational complexity relative to the unmodified one. By using a suitable threshold, it is possible to increase the tracking capability of the MCLMS algorithm compared to the LMS algorithm, but this causes slower convergence. Computer simulations confirm the mathematical analysis presented.

  16. 12 CFR 23.21 - Estimated residual value.

    Science.gov (United States)

    2010-01-01

    ... 12 Banks and Banking 1 2010-01-01 2010-01-01 false Estimated residual value. 23.21 Section 23.21...) Leases § 23.21 Estimated residual value. (a) Recovery of investment and costs. A national bank's estimate of the residual value of the property that the bank relies upon to satisfy the requirements of a full...

  17. Supercomputer implementation of finite element algorithms for high speed compressible flows. Progress report, period ending 30 June 1986

    International Nuclear Information System (INIS)

    Thornton, E.A.; Ramakrishnan, R.

    1986-06-01

    Prediction of compressible flow phenomena using the finite element method is of recent origin and considerable interest. Two shock capturing finite element formulations for high speed compressible flows are described. A Taylor-Galerkin formulation uses a Taylor series expansion in time coupled with a Galerkin weighted residual statement. The Taylor-Galerkin algorithms use explicit artificial dissipation, and the performance of three dissipation models are compared. A Petrov-Galerkin algorithm has as its basis the concepts of streamline upwinding. Vectorization strategies are developed to implement the finite element formulations on the NASA Langley VPS-32. The vectorization scheme results in finite element programs that use vectors of length of the order of the number of nodes or elements. The use of the vectorization procedure speeds up processing rates by over two orders of magnitude. The Taylor-Galerkin and Petrov-Galerkin algorithms are evaluated for 2D inviscid flows on criteria such as solution accuracy, shock resolution, computational speed and storage requirements. The convergence rates for both algorithms are enhanced by local time-stepping schemes. Extension of the vectorization procedure for predicting 2D viscous and 3D inviscid flows are demonstrated. Conclusions are drawn regarding the applicability of the finite element procedures for realistic problems that require hundreds of thousands of nodes

  18. Alcohol biodiesel from frying oil residues; Biodiesel etilico a partir de oleo de fritura residual

    Energy Technology Data Exchange (ETDEWEB)

    Festa, Brunna Simoes; Marques, Luiz Guilherme da Costa [Universidade Federal do Rio de Janeiro (IVIG/COPPE/UFRJ), RJ (Brazil). Coordenacao dos Programas de Pos-Graduacao de Engenharia. Inst. Virtual Internacional de Mudancas Globais], E-mail: lguilherme@ivig.coppe.ufrj.br

    2010-07-01

    This paper describes the reaction optimization and production of biodiesel through the use of frying residual oil made available by the restaurant placed at the PETROBRAS Research Center (CENPES-RJ), using ethanol, so that to permit the production of sustainable bio diesel. The environmental gains obtained by the utilization of residual oil, avoiding that this oil be released in the nature, and the economic gains coming from the generation and utilization of ethanol allowing the production of biodiesel be an viable alternative. The obtained results during laboratory tests shown that biodiesel produced from the transesterification in alkaline medium, of the frying residual oil collected presented a reaction yield of approximately 80% considering in mass.

  19. Algorithms as fetish: Faith and possibility in algorithmic work

    Directory of Open Access Journals (Sweden)

    Suzanne L Thomas

    2018-01-01

    Full Text Available Algorithms are powerful because we invest in them the power to do things. With such promise, they can transform the ordinary, say snapshots along a robotic vacuum cleaner’s route, into something much more, such as a clean home. Echoing David Graeber’s revision of fetishism, we argue that this easy slip from technical capabilities to broader claims betrays not the “magic” of algorithms but rather the dynamics of their exchange. Fetishes are not indicators of false thinking, but social contracts in material form. They mediate emerging distributions of power often too nascent, too slippery or too disconcerting to directly acknowledge. Drawing primarily on 2016 ethnographic research with computer vision professionals, we show how faith in what algorithms can do shapes the social encounters and exchanges of their production. By analyzing algorithms through the lens of fetishism, we can see the social and economic investment in some people’s labor over others. We also see everyday opportunities for social creativity and change. We conclude that what is problematic about algorithms is not their fetishization but instead their stabilization into full-fledged gods and demons – the more deserving objects of critique.

  20. A parallel electrostatic Particle-in-Cell method on unstructured tetrahedral grids for large-scale bounded collisionless plasma simulations

    Science.gov (United States)

    Averkin, Sergey N.; Gatsonis, Nikolaos A.

    2018-06-01

    An unstructured electrostatic Particle-In-Cell (EUPIC) method is developed on arbitrary tetrahedral grids for simulation of plasmas bounded by arbitrary geometries. The electric potential in EUPIC is obtained on cell vertices from a finite volume Multi-Point Flux Approximation of Gauss' law using the indirect dual cell with Dirichlet, Neumann and external circuit boundary conditions. The resulting matrix equation for the nodal potential is solved with a restarted generalized minimal residual method (GMRES) and an ILU(0) preconditioner algorithm, parallelized using a combination of node coloring and level scheduling approaches. The electric field on vertices is obtained using the gradient theorem applied to the indirect dual cell. The algorithms for injection, particle loading, particle motion, and particle tracking are parallelized for unstructured tetrahedral grids. The algorithms for the potential solver, electric field evaluation, loading, scatter-gather algorithms are verified using analytic solutions for test cases subject to Laplace and Poisson equations. Grid sensitivity analysis examines the L2 and L∞ norms of the relative error in potential, field, and charge density as a function of edge-averaged and volume-averaged cell size. Analysis shows second order of convergence for the potential and first order of convergence for the electric field and charge density. Temporal sensitivity analysis is performed and the momentum and energy conservation properties of the particle integrators in EUPIC are examined. The effects of cell size and timestep on heating, slowing-down and the deflection times are quantified. The heating, slowing-down and the deflection times are found to be almost linearly dependent on number of particles per cell. EUPIC simulations of current collection by cylindrical Langmuir probes in collisionless plasmas show good comparison with previous experimentally validated numerical results. These simulations were also used in a parallelization

  1. Quick fuzzy backpropagation algorithm.

    Science.gov (United States)

    Nikov, A; Stoeva, S

    2001-03-01

    A modification of the fuzzy backpropagation (FBP) algorithm called QuickFBP algorithm is proposed, where the computation of the net function is significantly quicker. It is proved that the FBP algorithm is of exponential time complexity, while the QuickFBP algorithm is of polynomial time complexity. Convergence conditions of the QuickFBP, resp. the FBP algorithm are defined and proved for: (1) single output neural networks in case of training patterns with different targets; and (2) multiple output neural networks in case of training patterns with equivalued target vector. They support the automation of the weights training process (quasi-unsupervised learning) establishing the target value(s) depending on the network's input values. In these cases the simulation results confirm the convergence of both algorithms. An example with a large-sized neural network illustrates the significantly greater training speed of the QuickFBP rather than the FBP algorithm. The adaptation of an interactive web system to users on the basis of the QuickFBP algorithm is presented. Since the QuickFBP algorithm ensures quasi-unsupervised learning, this implies its broad applicability in areas of adaptive and adaptable interactive systems, data mining, etc. applications.

  2. Tropospheric ozone column retrieval at northern mid-latitudes from the Ozone Monitoring Instrument by means of a neural network algorithm

    Directory of Open Access Journals (Sweden)

    P. Sellitto

    2011-11-01

    Full Text Available Monitoring tropospheric ozone from space is of critical importance in order to gain more thorough knowledge on phenomena affecting air quality and the greenhouse effect. Deriving information on tropospheric ozone from UV/VIS nadir satellite spectrometers is difficult owing to the weak sensitivity of the measured radiance spectra to variations of ozone in the troposphere. Here we propose an alternative method of analysis to retrieve tropospheric ozone columns from Ozone Monitoring Instrument radiances by means of a neural network algorithm. An extended set of ozone sonde measurements at northern mid-latitudes for the years 2004–2008 has been considered as the training and test data set. The design of the algorithm is extensively discussed. Our retrievals are compared to both tropospheric ozone residuals and optimal estimation retrievals over a similar independent test data set. Results show that our algorithm has comparable accuracy with respect to both correlative methods and its performance is slightly better over a subset containing only European ozone sonde stations. Possible sources of errors are analyzed. Finally, the capabilities of our algorithm to derive information on boundary layer ozone are studied and the results critically discussed.

  3. FINITE ELEMENT MODEL FOR PREDICTING RESIDUAL ...

    African Journals Online (AJOL)

    FINITE ELEMENT MODEL FOR PREDICTING RESIDUAL STRESSES IN ... the transverse residual stress in the x-direction (σx) had a maximum value of 375MPa ... the finite element method are in fair agreement with the experimental results.

  4. A New Modified Firefly Algorithm

    Directory of Open Access Journals (Sweden)

    Medha Gupta

    2016-07-01

    Full Text Available Nature inspired meta-heuristic algorithms studies the emergent collective intelligence of groups of simple agents. Firefly Algorithm is one of the new such swarm-based metaheuristic algorithm inspired by the flashing behavior of fireflies. The algorithm was first proposed in 2008 and since then has been successfully used for solving various optimization problems. In this work, we intend to propose a new modified version of Firefly algorithm (MoFA and later its performance is compared with the standard firefly algorithm along with various other meta-heuristic algorithms. Numerical studies and results demonstrate that the proposed algorithm is superior to existing algorithms.

  5. Quantum Computation and Algorithms

    International Nuclear Information System (INIS)

    Biham, O.; Biron, D.; Biham, E.; Grassi, M.; Lidar, D.A.

    1999-01-01

    It is now firmly established that quantum algorithms provide a substantial speedup over classical algorithms for a variety of problems, including the factorization of large numbers and the search for a marked element in an unsorted database. In this talk I will review the principles of quantum algorithms, the basic quantum gates and their operation. The combination of superposition and interference, that makes these algorithms efficient, will be discussed. In particular, Grover's search algorithm will be presented as an example. I will show that the time evolution of the amplitudes in Grover's algorithm can be found exactly using recursion equations, for any initial amplitude distribution

  6. Residue-Specific Side-Chain Polymorphisms via Particle Belief Propagation.

    Science.gov (United States)

    Ghoraie, Laleh Soltan; Burkowski, Forbes; Li, Shuai Cheng; Zhu, Mu

    2014-01-01

    Protein side chains populate diverse conformational ensembles in crystals. Despite much evidence that there is widespread conformational polymorphism in protein side chains, most of the X-ray crystallography data are modeled by single conformations in the Protein Data Bank. The ability to extract or to predict these conformational polymorphisms is of crucial importance, as it facilitates deeper understanding of protein dynamics and functionality. In this paper, we describe a computational strategy capable of predicting side-chain polymorphisms. Our approach extends a particular class of algorithms for side-chain prediction by modeling the side-chain dihedral angles more appropriately as continuous rather than discrete variables. Employing a new inferential technique known as particle belief propagation, we predict residue-specific distributions that encode information about side-chain polymorphisms. Our predicted polymorphisms are in relatively close agreement with results from a state-of-the-art approach based on X-ray crystallography data, which characterizes the conformational polymorphisms of side chains using electron density information, and has successfully discovered previously unmodeled conformations.

  7. Evaluation of heavy metals in hazardous automobile shredder residue thermal residue and immobilization with novel nano-size calcium dispersed reagent

    International Nuclear Information System (INIS)

    Lee, Chi-Hyeon; Truc, Nguyen Thi Thanh; Lee, Byeong-Kyu; Mitoma, Yoshiharu; Mallampati, Srinivasa Reddy

    2015-01-01

    Graphical abstract: Schematic representation of possible mechanisms determining the heavy metals immobilization efficiencyof ASR dust/thermal residues after treatment with nanometallic Ca/CaO/PO 4 . - Highlights: • Nanometallic Ca/CaO/PO 4 for heavy metals immobilization in ASR residue. • Heavy metals immobilization in dry condition attained about 95–100%. • Remaining heavy metals were lower than the Korean standard regulatory limit. • The amounts of heavy metals detectable on the ASR dust surface decreased. • Nanometallic Ca/CaO/PO 4 has a promising potential for heavy metal remediation. - Abstract: This study was conducted to synthesize and apply a nano-size calcium dispersed reagent as an immobilization material for heavy metal-contaminated automobile shredder residues (ASR) dust/thermal residues in dry condition. Simple mixing with a nanometallic Ca/CaO/PO 4 dispersion mixture immobilized 95–100% of heavy metals in ASR dust/thermal residues (including bottom ash, cavity ash, boiler and bag filter ash). The quantity of heavy metals leached from thermal residues after treatment by nanometallic Ca/CaO/PO 4 was lower than the Korean standard regulatory limit for hazardous waste landfills. The morphology and elemental composition of the nanometallic Ca/CaO-treated ASR residue were characterized by field emission scanning election microscopy combined with electron dispersive spectroscopy (FE-SEM/EDS). The results indicated that the amounts of heavy metals detectable on the ASR thermal residue surface decreased and the Ca/PO 4 mass percent increased. X-ray diffraction (XRD) pattern analysis indicated that the main fraction of enclosed/bound materials on ASR residue included Ca/PO 4 − associated crystalline complexes, and that immobile Ca/PO 4 salts remarkably inhibited the desorption of heavy metals from ASR residues. These results support the potential use of nanometallic Ca/CaO/PO 4 as a simple, suitable and highly efficient material for the gentle

  8. Evaluation of heavy metals in hazardous automobile shredder residue thermal residue and immobilization with novel nano-size calcium dispersed reagent

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Chi-Hyeon; Truc, Nguyen Thi Thanh [Department of Civil and Environmental Engineering, University of Ulsan, Daehak-ro 93, Nam-gu, Ulsan 680-749 (Korea, Republic of); Lee, Byeong-Kyu, E-mail: bklee@ulsan.ac.kr [Department of Civil and Environmental Engineering, University of Ulsan, Daehak-ro 93, Nam-gu, Ulsan 680-749 (Korea, Republic of); Mitoma, Yoshiharu [Department of Environmental Sciences, Prefectural University of Hiroshima, 562 Nanatsuka-Cho Shobara City, Hiroshima 727-0023 (Japan); Mallampati, Srinivasa Reddy, E-mail: srireddys@ulsan.ac.kr [Department of Civil and Environmental Engineering, University of Ulsan, Daehak-ro 93, Nam-gu, Ulsan 680-749 (Korea, Republic of)

    2015-10-15

    Graphical abstract: Schematic representation of possible mechanisms determining the heavy metals immobilization efficiencyof ASR dust/thermal residues after treatment with nanometallic Ca/CaO/PO{sub 4}. - Highlights: • Nanometallic Ca/CaO/PO{sub 4} for heavy metals immobilization in ASR residue. • Heavy metals immobilization in dry condition attained about 95–100%. • Remaining heavy metals were lower than the Korean standard regulatory limit. • The amounts of heavy metals detectable on the ASR dust surface decreased. • Nanometallic Ca/CaO/PO{sub 4} has a promising potential for heavy metal remediation. - Abstract: This study was conducted to synthesize and apply a nano-size calcium dispersed reagent as an immobilization material for heavy metal-contaminated automobile shredder residues (ASR) dust/thermal residues in dry condition. Simple mixing with a nanometallic Ca/CaO/PO{sub 4} dispersion mixture immobilized 95–100% of heavy metals in ASR dust/thermal residues (including bottom ash, cavity ash, boiler and bag filter ash). The quantity of heavy metals leached from thermal residues after treatment by nanometallic Ca/CaO/PO{sub 4} was lower than the Korean standard regulatory limit for hazardous waste landfills. The morphology and elemental composition of the nanometallic Ca/CaO-treated ASR residue were characterized by field emission scanning election microscopy combined with electron dispersive spectroscopy (FE-SEM/EDS). The results indicated that the amounts of heavy metals detectable on the ASR thermal residue surface decreased and the Ca/PO{sub 4} mass percent increased. X-ray diffraction (XRD) pattern analysis indicated that the main fraction of enclosed/bound materials on ASR residue included Ca/PO{sub 4}− associated crystalline complexes, and that immobile Ca/PO{sub 4} salts remarkably inhibited the desorption of heavy metals from ASR residues. These results support the potential use of nanometallic Ca/CaO/PO{sub 4} as a simple, suitable and

  9. Residual stress analysis in BWR pressure vessel attachments

    International Nuclear Information System (INIS)

    Dexter, R.J.; Leung, C.P.; Pont, D.

    1992-06-01

    Residual stresses from welding processes can be the primary driving force for stress corrosion cracking (SCC) in BWR components. Thus, a better understanding of the causes and nature of these residual stresses can help assess and remedy SCC. Numerical welding simulation software, such as SYSWELD, and material property data have been used to quantify residual stresses for application to SCC assessments in BWR components. Furthermore, parametric studies using SYSWELD have revealed which variables significantly affect predicted residual stress. Overall, numerical modeling techniques can be used to evaluate residual stress for SCC assessments of BWR components and to identify and plan future SCC research

  10. Semioptimal practicable algorithmic cooling

    International Nuclear Information System (INIS)

    Elias, Yuval; Mor, Tal; Weinstein, Yossi

    2011-01-01

    Algorithmic cooling (AC) of spins applies entropy manipulation algorithms in open spin systems in order to cool spins far beyond Shannon's entropy bound. Algorithmic cooling of nuclear spins was demonstrated experimentally and may contribute to nuclear magnetic resonance spectroscopy. Several cooling algorithms were suggested in recent years, including practicable algorithmic cooling (PAC) and exhaustive AC. Practicable algorithms have simple implementations, yet their level of cooling is far from optimal; exhaustive algorithms, on the other hand, cool much better, and some even reach (asymptotically) an optimal level of cooling, but they are not practicable. We introduce here semioptimal practicable AC (SOPAC), wherein a few cycles (typically two to six) are performed at each recursive level. Two classes of SOPAC algorithms are proposed and analyzed. Both attain cooling levels significantly better than PAC and are much more efficient than the exhaustive algorithms. These algorithms are shown to bridge the gap between PAC and exhaustive AC. In addition, we calculated the number of spins required by SOPAC in order to purify qubits for quantum computation. As few as 12 and 7 spins are required (in an ideal scenario) to yield a mildly pure spin (60% polarized) from initial polarizations of 1% and 10%, respectively. In the latter case, about five more spins are sufficient to produce a highly pure spin (99.99% polarized), which could be relevant for fault-tolerant quantum computing.

  11. Residual entanglement and sudden death: A direct connection

    International Nuclear Information System (INIS)

    Oliveira, J.G.G. de; Peixoto de Faria, J.G.; Nemes, M.C.

    2011-01-01

    We explore the results of [V. Coffman, et al., Phys. Rev. A 61 (2000) 052306] derived for general tripartite states in a dynamical context. We study a class of physically motivated tripartite systems. We show that whenever entanglement sudden death occurs in one of the partitions residual entanglement will appear. For fourpartite systems however, the appearance of residual entanglement is not conditioned by sudden death of entanglement. We can only say that if sudden death of entanglement occurs in some partition there will certainly be residual entanglement. -- Highlights: ► For tripartite systems we show there exists residual entanglement if sudden death occurs. ► For fourpartite systems, the residual entanglement is not conditioned by sudden death. ► If sudden death of entanglement occurs there will certainly be residual entanglement.

  12. Design of specimen for weld residual stress simulation

    International Nuclear Information System (INIS)

    Kim, Jin Weon; Park, Jong Sun; Lee, Kyung Soo

    2008-01-01

    The objective of this study is to design a laboratory specimen for simulating residual stress of circumferential butt welding of pipe. Specimen type and method for residual stress generation were proposed based on the review of prior studies and parametric finite element simulation. To prove the proposed specimen type and loading method, the residual stress was generated using the designed specimen by applying proposed method and was measured. The measured residual stress using X-ray diffraction reasonably agreed with the results of finite element simulation considered in the specimen design. Comparison of residual strains measured at several locations of specimen and given by finite element simulation also showed good agreement. Therefore, it is indicated that the designed specimen can reasonably simulate the residual stress of circumferential butt welding of pipe

  13. An Ordering Linear Unification Algorithm

    Institute of Scientific and Technical Information of China (English)

    胡运发

    1989-01-01

    In this paper,we present an ordering linear unification algorithm(OLU).A new idea on substituteion of the binding terms is introduced to the algorithm,which is able to overcome some drawbacks of other algorithms,e.g.,MM algorithm[1],RG1 and RG2 algorithms[2],Particularly,if we use the directed eyclie graphs,the algoritm needs not check the binding order,then the OLU algorithm can also be aplied to the infinite tree data struceture,and a higher efficiency can be expected.The paper focuses upon the discussion of OLU algorithm and a partial order structure with respect to the unification algorithm.This algorithm has been implemented in the GKD-PROLOG/VAX 780 interpreting system.Experimental results have shown that the algorithm is very simple and efficient.

  14. Residues in food derived from animals

    International Nuclear Information System (INIS)

    Grossklaus, D.

    1989-01-01

    The first chapter presents a survey of fundamentals and methods of the detection and analysis of residues in food derived from animals, also referring to the resulting health hazards to man, and to the relevant legal provisions. The subsequent chapters have been written by experts of the Federal Health Office, each dealing with particular types of residues such as those of veterinary drugs, additives to animal feeds, pesticide residues, and with environmental pollutants and the contamination of animal products with radionuclides. (MG) With 35 figs., 61 tabs [de

  15. Statistical Characterization of the Charge State and Residue Dependence of Low-Energy CID Peptide Dissociation Patterns

    International Nuclear Information System (INIS)

    Huang, Yingying; Triscari, Joseph M.; Tseng, George C.; Pasa-Tolic, Ljiljana; Lipton, Mary S.; Smith, Richard D.; Wysocki, Vicki H.

    2005-01-01

    Data mining was performed on 28 330 unique peptide tandem mass spectra for which sequences were assigned with high confidence. By dividing the spectra into different sets based on structural features and charge states of the corresponding peptides, chemical interactions involved in promoting specific cleavage patterns in gas-phase peptides were characterized. Pairwise fragmentation maps describing cleavages at all Xxx-Zzz residue combinations for b and y ions reveal that the difference in basicity between Arg and Lys results in different dissociation patterns for singly charged Arg- and Lys-ending tryptic peptides. While one dominant protonation form (proton localized) exists for Arg-ending peptides, a heterogeneous population of different protonated forms or more facile interconversion of protonated forms (proton partially mobile) exists for Lys-ending peptides. Cleavage C-terminal to acidic residues dominates spectra from peptides that have a localized proton and cleavage N-terminal to Pro dominates those that have a mobile or partially mobile proton. When Pro is absent from peptides that have a mobile or partially mobile proton, cleavage at each peptide bond becomes much more prominent. Whether the above patterns can be found in b ions, y ions, or both depends on the location of the proton holder(s). Enhanced cleavages C-terminal to branched aliphatic residues (Ile, Val, Leu) are observed in both b and y ions from peptides that have a mobile proton, as well as in y ions from peptides that have a partially mobile proton; enhanced cleavages N-terminal to these residues are observed in b ions from peptides that have a partially mobile proton. Statistical tools have been designed to visualize the fragmentation maps and measure the similarity between them. The pairwise cleavage patterns observed expand our knowledge of peptide gas-phase fragmentation behaviors and should be useful in algorithm development that employs improved models to predict fragment ion

  16. Measurement of residual stresses using fracture mechanics weight functions

    International Nuclear Information System (INIS)

    Fan, Y.

    2000-01-01

    A residual stress measurement method has been developed to quantify through-the-thickness residual stresses. Accurate measurement of residual stresses is crucial for many engineering structures. Fabrication processes such as welding and machining generate residual stresses that are difficult to predict. Residual stresses affect the integrity of structures through promoting failures due to brittle fracture, fatigue, stress corrosion cracking, and wear. In this work, the weight function theory of fracture mechanics is used to measure residual stresses. The weight function theory is an important development in computational fracture mechanics. Stress intensity factors for arbitrary stress distribution on the crack faces can be accurately and efficiently computed for predicting crack growth. This paper demonstrates that the weight functions are equally useful in measuring residual stresses. In this method, an artificial crack is created by a thin cut in a structure containing residual stresses. The cut relieves the residual stresses normal to the crack-face and allows the relieved residual stresses to deform the structure. Strain gages placed adjacent to the cut measure the relieved strains corresponding to incrementally increasing depths of the cut. The weight functions of the cracked body relate the measured strains to the residual stresses normal to the cut within the structure. The procedure details, such as numerical integration of the singular functions in applying the weight function method, will be discussed

  17. Measurement of residual stresses using fracture mechanics weight functions

    International Nuclear Information System (INIS)

    Fan, Y.

    2001-01-01

    A residual stress measurement method has been developed to quantify through-the-thickness residual stresses. Accurate measurement of residual stresses is crucial for many engineering structures. Fabrication processes such as welding and machining generate residual stresses that are difficult to predict. Residual stresses affect the integrity of structures through promoting failures due to brittle fracture, fatigue, stress corrosion cracking, and wear. In this work, the weight function theory of fracture mechanics is used to measure residual stresses. The weight function theory is an important development in computational fracture mechanics. Stress intensity factors for arbitrary stress distribution on the crack faces can be accurately and efficiently computed for predicting crack growth. This paper demonstrates that the weight functions are equally useful in measuring residual stresses. In this method, an artificial crack is created by a thin cut in a structure containing residual stresses. The cut relieves the residual stresses normal to the crack-face and allows the relieved residual stresses to deform the structure. Strain gages placed adjacent to the cut measure the relieved strains corresponding to incrementally increasing depths of the cut. The weight functions of the cracked body relate the measured strains to the residual stresses normal to the cut within the structure. The procedure details, such as numerical integration of the singular functions in applying the weight function method, will be discussed. (author)

  18. Effect of Residue Nitrogen Concentration and Time Duration on Carbon Mineralization Rate of Alfalfa Residues in Regions with Different Climatic Conditions

    Directory of Open Access Journals (Sweden)

    saeid shafiei

    2017-08-01

    Full Text Available Introduction Various factors like climatic conditions, vegetation, soil properties, topography, time, plant residue quality and crop management strategies affect the decomposition rate of organic carbon (OC and its residence time in soil. Plant residue management concerns nutrients recycling, carbon recycling in ecosystems and the increasing CO2 concentration in the atmosphere. Plant residue decomposition is a fundamental process in recycling of organic matter and elements in most ecosystems. Soil management, particularly plant residue management, changes soil organic matter both qualitatively and quantitatively. Soil respiration and carbon loss are affected by soil temperature, soil moisture, air temperature, solar radiation and precipitation. In natural agro-ecosystems, residue contains different concentrations of nitrogen. It is important to understand the rate and processes involved in plant residue decomposition, as these residues continue to be added to the soil under different weather conditions, especially in arid and semi-arid climates. Material and methods Organic carbon mineralization of alfalfa residue with different nitrogen concentrations was assessed in different climatic conditions using split-plot experiments over time and the effects of climate was determined using composite analysis. The climatic conditions were classified as warm-arid (Jiroft, temperate arid (Narab and cold semi-arid (Sardouiyeh using cluster analysis and the nitrogen (N concentrations of alfalfa residue were low, medium and high. The alfalfa residue incubated for four different time periods (2, 4, 6 and 8 months. The dynamics of organic carbon in different regions measured using litter bags (20×10 cm containing 20 g alfalfa residue of 2-10 mm length which were placed on the soil surface. Results and discussion The results of this study showed that in a warm-arid (Jiroft, carbon loss and the carbon decomposition rate constant were low in a cold semi

  19. The Effect of Processing on 14C- Chlofenvinphos Residues in Maize Oil and Bioavailability of its Cake Residues on Rats

    International Nuclear Information System (INIS)

    Mahdy, F.; El-Maghraby, S.

    2008-01-01

    Maize seed obtained from 14 C-chlofenvinphos treated plants contained 0.12 % of the applied dose. The insecticide residues in crude oil, methanol and coke amounted to 10 %, 6 % and 69 %, respectively of original residues inside the seeds.The 14 C activity in the crude oil could be a gradual reduced by the refining processes. The alkali treatment and bleaching steps are more effective steps in the refining processes remove about (63 %). The refined oil contained only about 17 % of the 14 C-residues originally present. The major residues in processed oil contain parent compound, in addition to five metabolites of the insecticide. When rats fed the extracted seeds (cake), the bound residues were found to be considerably bioavailable. After feeding rats for 5 days with the cake, a substantial amount of 14 C-residues was eliminated in the urine (59.5 %), while about 20 % was excreted in the feces. About 15 % of the radioactivity was distribution among various organs

  20. Computational prediction of protein hot spot residues.

    Science.gov (United States)

    Morrow, John Kenneth; Zhang, Shuxing

    2012-01-01

    Most biological processes involve multiple proteins interacting with each other. It has been recently discovered that certain residues in these protein-protein interactions, which are called hot spots, contribute more significantly to binding affinity than others. Hot spot residues have unique and diverse energetic properties that make them challenging yet important targets in the modulation of protein-protein complexes. Design of therapeutic agents that interact with hot spot residues has proven to be a valid methodology in disrupting unwanted protein-protein interactions. Using biological methods to determine which residues are hot spots can be costly and time consuming. Recent advances in computational approaches to predict hot spots have incorporated a myriad of features, and have shown increasing predictive successes. Here we review the state of knowledge around protein-protein interactions, hot spots, and give an overview of multiple in silico prediction techniques of hot spot residues.

  1. Stability analysis of a deterministic dose calculation for MRI-guided radiotherapy

    Science.gov (United States)

    Zelyak, O.; Fallone, B. G.; St-Aubin, J.

    2018-01-01

    Modern effort in radiotherapy to address the challenges of tumor localization and motion has led to the development of MRI guided radiotherapy technologies. Accurate dose calculations must properly account for the effects of the MRI magnetic fields. Previous work has investigated the accuracy of a deterministic linear Boltzmann transport equation (LBTE) solver that includes magnetic field, but not the stability of the iterative solution method. In this work, we perform a stability analysis of this deterministic algorithm including an investigation of the convergence rate dependencies on the magnetic field, material density, energy, and anisotropy expansion. The iterative convergence rate of the continuous and discretized LBTE including magnetic fields is determined by analyzing the spectral radius using Fourier analysis for the stationary source iteration (SI) scheme. The spectral radius is calculated when the magnetic field is included (1) as a part of the iteration source, and (2) inside the streaming-collision operator. The non-stationary Krylov subspace solver GMRES is also investigated as a potential method to accelerate the iterative convergence, and an angular parallel computing methodology is investigated as a method to enhance the efficiency of the calculation. SI is found to be unstable when the magnetic field is part of the iteration source, but unconditionally stable when the magnetic field is included in the streaming-collision operator. The discretized LBTE with magnetic fields using a space-angle upwind stabilized discontinuous finite element method (DFEM) was also found to be unconditionally stable, but the spectral radius rapidly reaches unity for very low-density media and increasing magnetic field strengths indicating arbitrarily slow convergence rates. However, GMRES is shown to significantly accelerate the DFEM convergence rate showing only a weak dependence on the magnetic field. In addition, the use of an angular parallel computing strategy

  2. Stability analysis of a deterministic dose calculation for MRI-guided radiotherapy.

    Science.gov (United States)

    Zelyak, O; Fallone, B G; St-Aubin, J

    2017-12-14

    Modern effort in radiotherapy to address the challenges of tumor localization and motion has led to the development of MRI guided radiotherapy technologies. Accurate dose calculations must properly account for the effects of the MRI magnetic fields. Previous work has investigated the accuracy of a deterministic linear Boltzmann transport equation (LBTE) solver that includes magnetic field, but not the stability of the iterative solution method. In this work, we perform a stability analysis of this deterministic algorithm including an investigation of the convergence rate dependencies on the magnetic field, material density, energy, and anisotropy expansion. The iterative convergence rate of the continuous and discretized LBTE including magnetic fields is determined by analyzing the spectral radius using Fourier analysis for the stationary source iteration (SI) scheme. The spectral radius is calculated when the magnetic field is included (1) as a part of the iteration source, and (2) inside the streaming-collision operator. The non-stationary Krylov subspace solver GMRES is also investigated as a potential method to accelerate the iterative convergence, and an angular parallel computing methodology is investigated as a method to enhance the efficiency of the calculation. SI is found to be unstable when the magnetic field is part of the iteration source, but unconditionally stable when the magnetic field is included in the streaming-collision operator. The discretized LBTE with magnetic fields using a space-angle upwind stabilized discontinuous finite element method (DFEM) was also found to be unconditionally stable, but the spectral radius rapidly reaches unity for very low-density media and increasing magnetic field strengths indicating arbitrarily slow convergence rates. However, GMRES is shown to significantly accelerate the DFEM convergence rate showing only a weak dependence on the magnetic field. In addition, the use of an angular parallel computing strategy

  3. On the Preconditioning of a Newton-Krylov Solver for a High-Order reconstructed Discontinuous Galerkin Discretization of All-Speed Compressible Flow with Phase Change for Application in Laser-Based Additive Manufacturing

    Energy Technology Data Exchange (ETDEWEB)

    Weston, Brian T. [Univ. of California, Davis, CA (United States)

    2017-05-17

    This dissertation focuses on the development of a fully-implicit, high-order compressible ow solver with phase change. The work is motivated by laser-induced phase change applications, particularly by the need to develop large-scale multi-physics simulations of the selective laser melting (SLM) process in metal additive manufacturing (3D printing). Simulations of the SLM process require precise tracking of multi-material solid-liquid-gas interfaces, due to laser-induced melting/ solidi cation and evaporation/condensation of metal powder in an ambient gas. These rapid density variations and phase change processes tightly couple the governing equations, requiring a fully compressible framework to robustly capture the rapid density variations of the ambient gas and the melting/evaporation of the metal powder. For non-isothermal phase change, the velocity is gradually suppressed through the mushy region by a variable viscosity and Darcy source term model. The governing equations are discretized up to 4th-order accuracy with our reconstructed Discontinuous Galerkin spatial discretization scheme and up to 5th-order accuracy with L-stable fully implicit time discretization schemes (BDF2 and ESDIRK3-5). The resulting set of non-linear equations is solved using a robust Newton-Krylov method, with the Jacobian-free version of the GMRES solver for linear iterations. Due to the sti nes associated with the acoustic waves and thermal and viscous/material strength e ects, preconditioning the GMRES solver is essential. A robust and scalable approximate block factorization preconditioner was developed, which utilizes the velocity-pressure (vP) and velocity-temperature (vT) Schur complement systems. This multigrid block reduction preconditioning technique converges for high CFL/Fourier numbers and exhibits excellent parallel and algorithmic scalability on classic benchmark problems in uid dynamics (lid-driven cavity ow and natural convection heat transfer) as well as for laser

  4. Residual number processing in dyscalculia ?

    OpenAIRE

    Cappelletti, Marinella; Price, Cathy J.

    2013-01-01

    Developmental dyscalculia – a congenital learning disability in understanding numerical concepts – is typically associated with parietal lobe abnormality. However, people with dyscalculia often retain some residual numerical abilities, reported in studies that otherwise focused on abnormalities in the dyscalculic brain. Here we took a different perspective by focusing on brain regions that support residual number processing in dyscalculia. All participants accurately performed semantic and ca...

  5. Americium recovery from reduction residues

    Science.gov (United States)

    Conner, W.V.; Proctor, S.G.

    1973-12-25

    A process for separation and recovery of americium values from container or bomb'' reduction residues comprising dissolving the residues in a suitable acid, adjusting the hydrogen ion concentration to a desired level by adding a base, precipitating the americium as americium oxalate by adding oxalic acid, digesting the solution, separating the precipitate, and thereafter calcining the americium oxalate precipitate to form americium oxide. (Official Gazette)

  6. VISUALIZATION OF PAGERANK ALGORITHM

    OpenAIRE

    Perhaj, Ervin

    2013-01-01

    The goal of the thesis is to develop a web application that help users understand the functioning of the PageRank algorithm. The thesis consists of two parts. First we develop an algorithm to calculate PageRank values of web pages. The input of algorithm is a list of web pages and links between them. The user enters the list through the web interface. From the data the algorithm calculates PageRank value for each page. The algorithm repeats the process, until the difference of PageRank va...

  7. RFID Location Algorithm

    Directory of Open Access Journals (Sweden)

    Wang Zi Min

    2016-01-01

    Full Text Available With the development of social services, people’s living standards improve further requirements, there is an urgent need for a way to adapt to the complex situation of the new positioning technology. In recent years, RFID technology have a wide range of applications in all aspects of life and production, such as logistics tracking, car alarm, security and other items. The use of RFID technology to locate, it is a new direction in the eyes of the various research institutions and scholars. RFID positioning technology system stability, the error is small and low-cost advantages of its location algorithm is the focus of this study.This article analyzes the layers of RFID technology targeting methods and algorithms. First, RFID common several basic methods are introduced; Secondly, higher accuracy to political network location method; Finally, LANDMARC algorithm will be described. Through this it can be seen that advanced and efficient algorithms play an important role in increasing RFID positioning accuracy aspects.Finally, the algorithm of RFID location technology are summarized, pointing out the deficiencies in the algorithm, and put forward a follow-up study of the requirements, the vision of a better future RFID positioning technology.

  8. A Uniform Energy Consumption Algorithm for Wireless Sensor and Actuator Networks Based on Dynamic Polling Point Selection

    Science.gov (United States)

    Li, Shuo; Peng, Jun; Liu, Weirong; Zhu, Zhengfa; Lin, Kuo-Chi

    2014-01-01

    Recent research has indicated that using the mobility of the actuator in wireless sensor and actuator networks (WSANs) to achieve mobile data collection can greatly increase the sensor network lifetime. However, mobile data collection may result in unacceptable collection delays in the network if the path of the actuator is too long. Because real-time network applications require meeting data collection delay constraints, planning the path of the actuator is a very important issue to balance the prolongation of the network lifetime and the reduction of the data collection delay. In this paper, a multi-hop routing mobile data collection algorithm is proposed based on dynamic polling point selection with delay constraints to address this issue. The algorithm can actively update the selection of the actuator's polling points according to the sensor nodes' residual energies and their locations while also considering the collection delay constraint. It also dynamically constructs the multi-hop routing trees rooted by these polling points to balance the sensor node energy consumption and the extension of the network lifetime. The effectiveness of the algorithm is validated by simulation. PMID:24451455

  9. Developing a NIR multispectral imaging for prediction and visualization of peanut protein content using variable selection algorithms

    Science.gov (United States)

    Cheng, Jun-Hu; Jin, Huali; Liu, Zhiwei

    2018-01-01

    The feasibility of developing a multispectral imaging method using important wavelengths from hyperspectral images selected by genetic algorithm (GA), successive projection algorithm (SPA) and regression coefficient (RC) methods for modeling and predicting protein content in peanut kernel was investigated for the first time. Partial least squares regression (PLSR) calibration model was established between the spectral data from the selected optimal wavelengths and the reference measured protein content ranged from 23.46% to 28.43%. The RC-PLSR model established using eight key wavelengths (1153, 1567, 1972, 2143, 2288, 2339, 2389 and 2446 nm) showed the best predictive results with the coefficient of determination of prediction (R2P) of 0.901, and root mean square error of prediction (RMSEP) of 0.108 and residual predictive deviation (RPD) of 2.32. Based on the obtained best model and image processing algorithms, the distribution maps of protein content were generated. The overall results of this study indicated that developing a rapid and online multispectral imaging system using the feature wavelengths and PLSR analysis is potential and feasible for determination of the protein content in peanut kernels.

  10. Improved multivariate polynomial factoring algorithm

    International Nuclear Information System (INIS)

    Wang, P.S.

    1978-01-01

    A new algorithm for factoring multivariate polynomials over the integers based on an algorithm by Wang and Rothschild is described. The new algorithm has improved strategies for dealing with the known problems of the original algorithm, namely, the leading coefficient problem, the bad-zero problem and the occurrence of extraneous factors. It has an algorithm for correctly predetermining leading coefficients of the factors. A new and efficient p-adic algorithm named EEZ is described. Bascially it is a linearly convergent variable-by-variable parallel construction. The improved algorithm is generally faster and requires less store then the original algorithm. Machine examples with comparative timing are included

  11. Governance by algorithms

    Directory of Open Access Journals (Sweden)

    Francesca Musiani

    2013-08-01

    Full Text Available Algorithms are increasingly often cited as one of the fundamental shaping devices of our daily, immersed-in-information existence. Their importance is acknowledged, their performance scrutinised in numerous contexts. Yet, a lot of what constitutes 'algorithms' beyond their broad definition as “encoded procedures for transforming input data into a desired output, based on specified calculations” (Gillespie, 2013 is often taken for granted. This article seeks to contribute to the discussion about 'what algorithms do' and in which ways they are artefacts of governance, providing two examples drawing from the internet and ICT realm: search engine queries and e-commerce websites’ recommendations to customers. The question of the relationship between algorithms and rules is likely to occupy an increasingly central role in the study and the practice of internet governance, in terms of both institutions’ regulation of algorithms, and algorithms’ regulation of our society.

  12. Corn Residue Use by Livestock in the United States

    Directory of Open Access Journals (Sweden)

    Marty R. Schmer

    2017-01-01

    Full Text Available Corn ( L. residue grazing or harvest provides a simple and economical practice to integrate crops and livestock, but limited information is available on how widespread corn residue utilization is practiced by US producers. In 2010, the USDA Economic Research Service surveyed producers from 19 states on corn grain and residue management practices. Total corn residue grazed or harvested was 4.87 million ha. Approximately 4.06 million ha was grazed by 11.7 million livestock (primarily cattle in 2010. The majority of grazed corn residue occurred in Nebraska (1.91 million ha, Iowa (385,000 ha, South Dakota (361,000 ha, and Kansas (344,000 ha. Average grazing days ranged from 10 to 73 d (mean = 40 d. Corn residue harvests predominantly occurred in the central and northern Corn Belt, with an estimated 2.9 Tg of corn residue harvested across the 19 states. This survey highlights the importance of corn residue for US livestock, particularly in the western Corn Belt.

  13. 40 CFR 180.586 - Clothianidin; tolerances for residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Clothianidin; tolerances for residues... § 180.586 Clothianidin; tolerances for residues. (a) General. Tolerances are established for residues of the insecticide clothianidin, including its metabolites and degradates. Compliance with the tolerance...

  14. Guidelines for selection and presentation of residue values of pesticides

    NARCIS (Netherlands)

    Velde-Koerts T van der; Hoeven-Arentzen PH van; Ossendorp BC; RIVM-SIR

    2004-01-01

    Pesticide residue assessments are executed to establish legal limits, called Maximum Residue Limits (MRLs). MRLs are derived from the results of these pesticide residue trials, which are performed according to critical Good Agricultural Practice. Only one residue value per residue trial may be

  15. Semantic Tagging with Deep Residual Networks

    NARCIS (Netherlands)

    Bjerva, Johannes; Plank, Barbara; Bos, Johan

    2016-01-01

    We propose a novel semantic tagging task, semtagging, tailored for the purpose of multilingual semantic parsing, and present the first tagger using deep residual networks (ResNets). Our tagger uses both word and character representations and includes a novel residual bypass architecture. We evaluate

  16. Algorithmic and user study of an autocompletion algorithm on a large medical vocabulary.

    Science.gov (United States)

    Sevenster, Merlijn; van Ommering, Rob; Qian, Yuechen

    2012-02-01

    Autocompletion supports human-computer interaction in software applications that let users enter textual data. We will be inspired by the use case in which medical professionals enter ontology concepts, catering the ongoing demand for structured and standardized data in medicine. Goal is to give an algorithmic analysis of one particular autocompletion algorithm, called multi-prefix matching algorithm, which suggests terms whose words' prefixes contain all words in the string typed by the user, e.g., in this sense, opt ner me matches optic nerve meningioma. Second we aim to investigate how well it supports users entering concepts from a large and comprehensive medical vocabulary (snomed ct). We give a concise description of the multi-prefix algorithm, and sketch how it can be optimized to meet required response time. Performance will be compared to a baseline algorithm, which gives suggestions that extend the string typed by the user to the right, e.g. optic nerve m gives optic nerve meningioma, but opt ner me does not. We conduct a user experiment in which 12 participants are invited to complete 40 snomed ct terms with the baseline algorithm and another set of 40 snomed ct terms with the multi-prefix algorithm. Our results show that users need significantly fewer keystrokes when supported by the multi-prefix algorithm than when supported by the baseline algorithm. The proposed algorithm is a competitive candidate for searching and retrieving terms from a large medical ontology. Copyright © 2011 Elsevier Inc. All rights reserved.

  17. Optimal Fungal Space Searching Algorithms.

    Science.gov (United States)

    Asenova, Elitsa; Lin, Hsin-Yu; Fu, Eileen; Nicolau, Dan V; Nicolau, Dan V

    2016-10-01

    Previous experiments have shown that fungi use an efficient natural algorithm for searching the space available for their growth in micro-confined networks, e.g., mazes. This natural "master" algorithm, which comprises two "slave" sub-algorithms, i.e., collision-induced branching and directional memory, has been shown to be more efficient than alternatives, with one, or the other, or both sub-algorithms turned off. In contrast, the present contribution compares the performance of the fungal natural algorithm against several standard artificial homologues. It was found that the space-searching fungal algorithm consistently outperforms uninformed algorithms, such as Depth-First-Search (DFS). Furthermore, while the natural algorithm is inferior to informed ones, such as A*, this under-performance does not importantly increase with the increase of the size of the maze. These findings suggest that a systematic effort of harvesting the natural space searching algorithms used by microorganisms is warranted and possibly overdue. These natural algorithms, if efficient, can be reverse-engineered for graph and tree search strategies.

  18. [Migrants from disposable gloves and residual acrylonitrile].

    Science.gov (United States)

    Wakui, C; Kawamura, Y; Maitani, T

    2001-10-01

    Disposable gloves made from polyvinyl chloride with and without di(2-ethylhexyl) phthalate (PVC-DEHP, PVC-NP), polyethylene (PE), natural rubber (NR) and nitrile-butadiene rubber (NBR) were investigated with respect to evaporation residue, migrated metals, migrants and residual acrylonitrile. The evaporation residue found in n-heptane was 870-1,300 ppm from PVC-DEHP and PVC-NP, which was due to the plasticizers. Most of the PE gloves had low evaporation residue levels and migrants, except for the glove designated as antibacterial, which released copper and zinc into 4% acetic acid. For the NR and NBR gloves, the evaporation residue found in 4% acetic acid was 29-180 ppm. They also released over 10 ppm of calcium and 6 ppm of zinc into 4% acetic acid, and 1.68-8.37 ppm of zinc di-ethyldithiocarbamate and zinc di-n-butyldithiocarbamate used as vulcanization accelerators into n-heptane. The acrylonitrile content was 0.40-0.94 ppm in NBR gloves.

  19. Harvesting and handling agricultural residues for energy

    Energy Technology Data Exchange (ETDEWEB)

    Jenkins, B.M.; Summer, H.R.

    1986-05-01

    Significant progress in understanding the needs for design of agricultural residue collection and handling systems has been made but additional research is required. Recommendations are made for research to (a) integrate residue collection and handling systems into general agricultural practices through the development of multi-use equipment and total harvest systems; (b) improve methods for routine evaluation of agricultural residue resources, possibly through remote sensing and image processing; (c) analyze biomass properties to obtain detailed data relevant to engineering design and analysis; (d) evaluate long-term environmental, social, and agronomic impacts of residue collection; (e) develop improved equipment with higher capacities to reduce residue collection and handling costs, with emphasis on optimal design of complete systems including collection, transportation, processing, storage, and utilization; and (f) produce standard forms of biomass fuels or products to enhance material handling and expand biomass markets through improved reliability and automatic control of biomass conversion and other utilization systems. 118 references.

  20. Protein structure based prediction of catalytic residues.

    Science.gov (United States)

    Fajardo, J Eduardo; Fiser, Andras

    2013-02-22

    Worldwide structural genomics projects continue to release new protein structures at an unprecedented pace, so far nearly 6000, but only about 60% of these proteins have any sort of functional annotation. We explored a range of features that can be used for the prediction of functional residues given a known three-dimensional structure. These features include various centrality measures of nodes in graphs of interacting residues: closeness, betweenness and page-rank centrality. We also analyzed the distance of functional amino acids to the general center of mass (GCM) of the structure, relative solvent accessibility (RSA), and the use of relative entropy as a measure of sequence conservation. From the selected features, neural networks were trained to identify catalytic residues. We found that using distance to the GCM together with amino acid type provide a good discriminant function, when combined independently with sequence conservation. Using an independent test set of 29 annotated protein structures, the method returned 411 of the initial 9262 residues as the most likely to be involved in function. The output 411 residues contain 70 of the annotated 111 catalytic residues. This represents an approximately 14-fold enrichment of catalytic residues on the entire input set (corresponding to a sensitivity of 63% and a precision of 17%), a performance competitive with that of other state-of-the-art methods. We found that several of the graph based measures utilize the same underlying feature of protein structures, which can be simply and more effectively captured with the distance to GCM definition. This also has the added the advantage of simplicity and easy implementation. Meanwhile sequence conservation remains by far the most influential feature in identifying functional residues. We also found that due the rapid changes in size and composition of sequence databases, conservation calculations must be recalibrated for specific reference databases.

  1. Identification of Key Residues for Enzymatic Carboxylate Reduction

    Directory of Open Access Journals (Sweden)

    Holly Stolterfoht

    2018-02-01

    Full Text Available Carboxylate reductases (CARs, E.C. 1.2.1.30 generate aldehydes from their corresponding carboxylic acid with high selectivity. Little is known about the structure of CARs and their catalytically important amino acid residues. The identification of key residues for carboxylate reduction provides a starting point to gain deeper understanding of enzymatic carboxylate reduction. A multiple sequence alignment of CARs with confirmed activity recently identified in our lab and from the literature revealed a fingerprint of conserved amino acids. We studied the function of conserved residues by multiple sequence alignments and mutational replacements of these residues. In this study, single-site alanine variants of Neurospora crassa CAR were investigated to determine the contribution of conserved residues to the function, expressability or stability of the enzyme. The effect of amino acid replacements was investigated by analyzing enzymatic activity of the variants in vivo and in vitro. Supported by molecular modeling, we interpreted that five of these residues are essential for catalytic activity, or substrate and co-substrate binding. We identified amino acid residues having significant impact on CAR activity. Replacement of His 237, Glu 433, Ser 595, Tyr 844, and Lys 848 by Ala abolish CAR activity, indicating their key role in acid reduction. These results may assist in the functional annotation of CAR coding genes in genomic databases. While some other conserved residues decreased activity or had no significant impact, four residues increased the specific activity of NcCAR variants when replaced by alanine. Finally, we showed that NcCAR wild-type and mutants efficiently reduce aliphatic acids.

  2. Extreme Scale FMM-Accelerated Boundary Integral Equation Solver for Wave Scattering

    KAUST Repository

    AbdulJabbar, Mustafa Abdulmajeed; Al Farhan, Mohammed; Al-Harthi, Noha A.; Chen, Rui; Yokota, Rio; Bagci, Hakan; Keyes, David E.

    2018-01-01

    scattering, which uses FMM as a matrix-vector multiplication inside the GMRES iterative method. Our FMM Helmholtz kernels treat nontrivial singular and near-field integration points. We implement highly optimized kernels for both shared and distributed memory

  3. Autonomous Star Tracker Algorithms

    DEFF Research Database (Denmark)

    Betto, Maurizio; Jørgensen, John Leif; Kilsgaard, Søren

    1998-01-01

    Proposal, in response to an ESA R.f.P., to design algorithms for autonomous star tracker operations.The proposal also included the development of a star tracker breadboard to test the algorithms performances.......Proposal, in response to an ESA R.f.P., to design algorithms for autonomous star tracker operations.The proposal also included the development of a star tracker breadboard to test the algorithms performances....

  4. Residual stress in spin-cast polyurethane thin films

    Energy Technology Data Exchange (ETDEWEB)

    Zhang, Hong; Zhang, Li, E-mail: lizhang@mae.cuhk.edu.hk [Department of Mechanical and Automation Engineering, The Chinese University of Hong Kong, Shatin N.T., Hong Kong (China); Chow Yuk Ho Technology Centre for Innovative Medicine, The Chinese University of Hong Kong, Shatin N.T., Hong Kong (China)

    2015-01-19

    Residual stress is inevitable during spin-casting. Herein, we report a straightforward method to evaluate the residual stress in as-cast polyurethane thin films using area shrinkage measurement of films in floating state, which shows that the residual stress is independent of radial location on the substrate and decreased with decreasing film thickness below a critical value. We demonstrate that the residual stress is developed due to the solvent evaporation after vitrification during spin-casting and the polymer chains in thin films may undergo vitrification at an increased concentration. The buildup of residual stress in spin-cast polymer films provides an insight into the size effects on the nature of polymer thin films.

  5. Residual stress studies of austenitic and ferritic steels

    International Nuclear Information System (INIS)

    Chrenko, R.M.

    1978-01-01

    Residual studies have been made on austenitic and ferritic steels of the types used as structural materials. The residual stress results presented here will include residual stress measurements in the heat-affected zone on butt welded Type 304 stainless steel pipes, and the stresses induced in Type 304 austenitic stainless steel and Type A508 ferritic steel by several surface preparations. Such surface preparation procedures as machining and grinding can induce large directionality effects in the residual stresses determined by X-ray techniques and some typical data will be presented. A brief description is given of the mobile X-ray residual stress apparatus used to obtain most of the data in these studies. (author)

  6. CAB-Align: A Flexible Protein Structure Alignment Method Based on the Residue-Residue Contact Area.

    Directory of Open Access Journals (Sweden)

    Genki Terashi

    Full Text Available Proteins are flexible, and this flexibility has an essential functional role. Flexibility can be observed in loop regions, rearrangements between secondary structure elements, and conformational changes between entire domains. However, most protein structure alignment methods treat protein structures as rigid bodies. Thus, these methods fail to identify the equivalences of residue pairs in regions with flexibility. In this study, we considered that the evolutionary relationship between proteins corresponds directly to the residue-residue physical contacts rather than the three-dimensional (3D coordinates of proteins. Thus, we developed a new protein structure alignment method, contact area-based alignment (CAB-align, which uses the residue-residue contact area to identify regions of similarity. The main purpose of CAB-align is to identify homologous relationships at the residue level between related protein structures. The CAB-align procedure comprises two main steps: First, a rigid-body alignment method based on local and global 3D structure superposition is employed to generate a sufficient number of initial alignments. Then, iterative dynamic programming is executed to find the optimal alignment. We evaluated the performance and advantages of CAB-align based on four main points: (1 agreement with the gold standard alignment, (2 alignment quality based on an evolutionary relationship without 3D coordinate superposition, (3 consistency of the multiple alignments, and (4 classification agreement with the gold standard classification. Comparisons of CAB-align with other state-of-the-art protein structure alignment methods (TM-align, FATCAT, and DaliLite using our benchmark dataset showed that CAB-align performed robustly in obtaining high-quality alignments and generating consistent multiple alignments with high coverage and accuracy rates, and it performed extremely well when discriminating between homologous and nonhomologous pairs of proteins

  7. CAB-Align: A Flexible Protein Structure Alignment Method Based on the Residue-Residue Contact Area.

    Science.gov (United States)

    Terashi, Genki; Takeda-Shitaka, Mayuko

    2015-01-01

    Proteins are flexible, and this flexibility has an essential functional role. Flexibility can be observed in loop regions, rearrangements between secondary structure elements, and conformational changes between entire domains. However, most protein structure alignment methods treat protein structures as rigid bodies. Thus, these methods fail to identify the equivalences of residue pairs in regions with flexibility. In this study, we considered that the evolutionary relationship between proteins corresponds directly to the residue-residue physical contacts rather than the three-dimensional (3D) coordinates of proteins. Thus, we developed a new protein structure alignment method, contact area-based alignment (CAB-align), which uses the residue-residue contact area to identify regions of similarity. The main purpose of CAB-align is to identify homologous relationships at the residue level between related protein structures. The CAB-align procedure comprises two main steps: First, a rigid-body alignment method based on local and global 3D structure superposition is employed to generate a sufficient number of initial alignments. Then, iterative dynamic programming is executed to find the optimal alignment. We evaluated the performance and advantages of CAB-align based on four main points: (1) agreement with the gold standard alignment, (2) alignment quality based on an evolutionary relationship without 3D coordinate superposition, (3) consistency of the multiple alignments, and (4) classification agreement with the gold standard classification. Comparisons of CAB-align with other state-of-the-art protein structure alignment methods (TM-align, FATCAT, and DaliLite) using our benchmark dataset showed that CAB-align performed robustly in obtaining high-quality alignments and generating consistent multiple alignments with high coverage and accuracy rates, and it performed extremely well when discriminating between homologous and nonhomologous pairs of proteins in both

  8. Calcination/dissolution residue treatment

    International Nuclear Information System (INIS)

    Knight, R.C.; Creed, R.F.; Patello, G.K.; Hollenberg, G.W.; Buehler, M.F.; O'Rourke, S.M.; Visnapuu, A.; McLaughlin, D.F.

    1994-09-01

    Currently, high-level wastes are stored underground in steel-lined tanks at the Hanford site. Current plans call for the chemical pretreatment of these wastes before their immobilization in stable glass waste forms. One candidate pretreatment approach, calcination/dissolution, performs an alkaline fusion of the waste and creates a high-level/low-level partition based on the aqueous solubilities of the components of the product calcine. Literature and laboratory studies were conducted with the goal of finding a residue treatment technology that would decrease the quantity of high-level waste glass required following calcination/dissolution waste processing. Four elements, Fe, Ni, Bi, and U, postulated to be present in the high-level residue fraction were identified as being key to the quantity of high-level glass formed. Laboratory tests of the candidate technologies with simulant high-level residues showed reductive roasting followed by carbonyl volatilization to be successful in removing Fe, Ni, and Bi. Subsequent bench-scale tests on residues from calcination/dissolution processing of genuine Hanford Site tank waste showed Fe was separated with radioelement decontamination factors of 70 to 1,000 times with respect to total alpha activity. Thermodynamic analyses of the calcination of five typical Hanford Site tank waste compositions also were performed. The analyses showed sodium hydroxide to be the sole molten component in the waste calcine and emphasized the requirement for waste blending if fluid calcines are to be achieved. Other calcine phases identified in the thermodynamic analysis indicate the significant thermal reconstitution accomplished in calcination

  9. Cover crop residue management for optimizing weed control

    NARCIS (Netherlands)

    Kruidhof, H.M.; Bastiaans, L.; Kropff, M.J.

    2009-01-01

    Although residue management seems a key factor in residue-mediated weed suppression, very few studies have systematically compared the influence of different residue management strategies on the establishment of crop and weed species. We evaluated the effect of several methods of pre-treatment and

  10. Decomposition of rice residue in tropical soils, 2

    International Nuclear Information System (INIS)

    Yoneyama, Tadakatsu; Yoshida, Tomio

    1977-01-01

    The decomposition processes of intact rice residue (leaf blades) in the Maahas soil in the Philippines were investigated. Three sets of beakers simulating both lowland and upland conditions were incubated in the dark at 30 deg. C. One set of beakers had neither rice residue nor fertilizer. Pieces of leaf blades weighing 204 mg (dry weight) were inserted in the second set. Pieces of leaf blades were inserted in the third set, and 200 ppm of fertilizer nitrogen as 15 N-labelled ammonium sulfate was added. The experiment dealt with the nitrogen immobilization by rice residue under lowland and upland conditions. The rice residue which has contained low nitrogen absorbed nitrogen from the soil and from the added fertilizer (ammonium sulfate) during its decomposition under both conditions. Under the lowland condition, the amount of nitrogen immobilized was small during the first week, but became large after 2 or 3 weeks. Under the upland condition, the immobilized nitrogen reached its maximum during the first week, but the amount was not so large as that under the lowland condition. The added fertilizer stimulated the decrease of weight of the rice residue in the early incubation period, but retarded it later under both conditions. The absorption of fertilizer by the rice residue ceased at the early stage of residue decomposition, but the nitrogen content of the residue continued to increase. (Iwakiri, K.)

  11. Cyolane residues in milk of lactating goats

    International Nuclear Information System (INIS)

    Zayed, S.M.A.D.; Osman, A.; Fakhr, I.M.I.

    1981-01-01

    Consecutive feeding of lactating goats with 14 C-alkyl labelled cyolane for 5 days at dietary levels 8 and 16 ppm resulted in the appearance of measurable insecticide residues in milk (0.02-0.04 mg/kg). The residue levels were markedly reduced after a withdrawal period of 7 days. Analysis of urine and milk residues showed the presence of similar metabolites in addition to the parent compound. The major part of the residue consisted of mono-, diethyl phosphate and 2 hydrophilic unknown metabolites. The erythrocyte cholinesterase activity was reduced to about 50% after 24 hours whereas the plasma enzyme was only slightly affected. The animals remained symptom-free during the experimental period. (author)

  12. Surgical treatment for residual or recurrent strabismus

    Directory of Open Access Journals (Sweden)

    Tao Wang

    2014-12-01

    Full Text Available Although the surgical treatment is a relatively effective and predictable method for correcting residual or recurrent strabismus, such as posterior fixation sutures, medial rectus marginal myotomy, unilateral or bilateral rectus re-recession and resection, unilateral lateral rectus recession and adjustable suture, no standard protocol is established for the surgical style. Different surgical approaches have been recommended for correcting residual or recurrent strabismus. The choice of the surgical procedure depends on the former operation pattern and the surgical dosages applied on the patients, residual or recurrent angle of deviation and the operator''s preference and experience. This review attempts to outline recent publications and current opinion in the management of residual or recurrent esotropia and exotropia.

  13. Fast geometric algorithms

    International Nuclear Information System (INIS)

    Noga, M.T.

    1984-01-01

    This thesis addresses a number of important problems that fall within the framework of the new discipline of Computational Geometry. The list of topics covered includes sorting and selection, convex hull algorithms, the L 1 hull, determination of the minimum encasing rectangle of a set of points, the Euclidean and L 1 diameter of a set of points, the metric traveling salesman problem, and finding the superrange of star-shaped and monotype polygons. The main theme of all the work was to develop a set of very fast state-of-the-art algorithms that supersede any rivals in terms of speed and ease of implementation. In some cases existing algorithms were refined; for others new techniques were developed that add to the present database of fast adaptive geometric algorithms. What emerges is a collection of techniques that is successful at merging modern tools developed in analysis of algorithms with those of classical geometry

  14. Residual stress analysis in reactor pressure vessel attachments

    International Nuclear Information System (INIS)

    Dexter, R.J.; Pont, D.

    1991-08-01

    Residual stresses in cladding and welded attachments could contribute to the problem of stress-corrosion cracking in boiling-water reactors (BWR). As part of a larger program aimed at quantifying residual stress in BWR components, models that would be applicable for predicting residual stress in BWR components are reviewed and documented. The review includes simple methods of estimating residual stresses as well as advanced finite-element software. In general, simple methods are capable of predicting peak magnitudes of residual stresses but are incapable of adequately characterizing the distribution of residual stresses. Ten groups of researchers using finite-element software are reviewed in detail. For each group, the assumptions of the model, possible simplifications, material property data, and specific applications are discussed. The most accurate results are obtained when a metallurgical simulation is performed, transformation plasticity effects are included, and the heating and cooling parts of the welding thermal cycle are simulated. Two models are identified which can provide these features. The present state of these models and the material property data available in the literature are adequate to quantify residual stress in BWR components

  15. [Development of residual voltage testing equipment].

    Science.gov (United States)

    Zeng, Xiaohui; Wu, Mingjun; Cao, Li; He, Jinyi; Deng, Zhensheng

    2014-07-01

    For the existing measurement methods of residual voltage which can't turn the power off at peak voltage exactly and simultaneously display waveforms, a new residual voltage detection method is put forward in this paper. First, the zero point of the power supply is detected with zero cross detection circuit and is inputted to a single-chip microcomputer in the form of pulse signal. Secend, when the zero point delays to the peak voltage, the single-chip microcomputer sends control signal to power off the relay. At last, the waveform of the residual voltage is displayed on a principal computer or oscilloscope. The experimental results show that the device designed in this paper can turn the power off at peak voltage and is able to accurately display the voltage waveform immediately after power off and the standard deviation of the residual voltage is less than 0.2 V at exactly one second and later.

  16. Genetic Algorithm Applied to the Eigenvalue Equalization Filtered-x LMS Algorithm (EE-FXLMS

    Directory of Open Access Journals (Sweden)

    Stephan P. Lovstedt

    2008-01-01

    Full Text Available The FXLMS algorithm, used extensively in active noise control (ANC, exhibits frequency-dependent convergence behavior. This leads to degraded performance for time-varying tonal noise and noise with multiple stationary tones. Previous work by the authors proposed the eigenvalue equalization filtered-x least mean squares (EE-FXLMS algorithm. For that algorithm, magnitude coefficients of the secondary path transfer function are modified to decrease variation in the eigenvalues of the filtered-x autocorrelation matrix, while preserving the phase, giving faster convergence and increasing overall attenuation. This paper revisits the EE-FXLMS algorithm, using a genetic algorithm to find magnitude coefficients that give the least variation in eigenvalues. This method overcomes some of the problems with implementing the EE-FXLMS algorithm arising from finite resolution of sampled systems. Experimental control results using the original secondary path model, and a modified secondary path model for both the previous implementation of EE-FXLMS and the genetic algorithm implementation are compared.

  17. Model-checking techniques based on cumulative residuals.

    Science.gov (United States)

    Lin, D Y; Wei, L J; Ying, Z

    2002-03-01

    Residuals have long been used for graphical and numerical examinations of the adequacy of regression models. Conventional residual analysis based on the plots of raw residuals or their smoothed curves is highly subjective, whereas most numerical goodness-of-fit tests provide little information about the nature of model misspecification. In this paper, we develop objective and informative model-checking techniques by taking the cumulative sums of residuals over certain coordinates (e.g., covariates or fitted values) or by considering some related aggregates of residuals, such as moving sums and moving averages. For a variety of statistical models and data structures, including generalized linear models with independent or dependent observations, the distributions of these stochastic processes tinder the assumed model can be approximated by the distributions of certain zero-mean Gaussian processes whose realizations can be easily generated by computer simulation. Each observed process can then be compared, both graphically and numerically, with a number of realizations from the Gaussian process. Such comparisons enable one to assess objectively whether a trend seen in a residual plot reflects model misspecification or natural variation. The proposed techniques are particularly useful in checking the functional form of a covariate and the link function. Illustrations with several medical studies are provided.

  18. Detecting organic gunpowder residues from handgun use

    Science.gov (United States)

    MacCrehan, William A.; Ricketts, K. Michelle; Baltzersen, Richard A.; Rowe, Walter F.

    1999-02-01

    The gunpowder residues that remain after the use of handguns or improvised explosive devices pose a challenge for the forensic investigator. Can these residues be reliably linked to a specific gunpowder or ammunition? We investigated the possibility by recovering and measuring the composition of organic additives in smokeless powder and its post-firing residues. By determining gunpowder additives such as nitroglycerin, dinitrotoluene, ethyl- and methylcentralite, and diphenylamine, we hope to identify the type of gunpowder in the residues and perhaps to provide evidence of a match to a sample of unfired powder. The gunpowder additives were extracted using an automated technique, pressurized fluid extraction (PFE). The conditions for the quantitative extraction of the additives using neat and solvent-modified supercritical carbon dioxide were investigated. All of the major gunpowder additives can be determined with baseline resolution using capillary electrophoresis (CE) with a micellar agent and UV absorbance detection. A study of candidate internal standards for use in the CE method is also presented. The PFE/CE technique is used to evaluate a new residue sampling protocol--asking shooters to blow their noses. In addition, an initial investigation of the compositional differences among unfired and post-fired .22 handgun residues is presented.

  19. The gradient boosting algorithm and random boosting for genome-assisted evaluation in large data sets.

    Science.gov (United States)

    González-Recio, O; Jiménez-Montero, J A; Alenda, R

    2013-01-01

    In the next few years, with the advent of high-density single nucleotide polymorphism (SNP) arrays and genome sequencing, genomic evaluation methods will need to deal with a large number of genetic variants and an increasing sample size. The boosting algorithm is a machine-learning technique that may alleviate the drawbacks of dealing with such large data sets. This algorithm combines different predictors in a sequential manner with some shrinkage on them; each predictor is applied consecutively to the residuals from the committee formed by the previous ones to form a final prediction based on a subset of covariates. Here, a detailed description is provided and examples using a toy data set are included. A modification of the algorithm called "random boosting" was proposed to increase predictive ability and decrease computation time of genome-assisted evaluation in large data sets. Random boosting uses a random selection of markers to add a subsequent weak learner to the predictive model. These modifications were applied to a real data set composed of 1,797 bulls genotyped for 39,714 SNP. Deregressed proofs of 4 yield traits and 1 type trait from January 2009 routine evaluations were used as dependent variables. A 2-fold cross-validation scenario was implemented. Sires born before 2005 were used as a training sample (1,576 and 1,562 for production and type traits, respectively), whereas younger sires were used as a testing sample to evaluate predictive ability of the algorithm on yet-to-be-observed phenotypes. Comparison with the original algorithm was provided. The predictive ability of the algorithm was measured as Pearson correlations between observed and predicted responses. Further, estimated bias was computed as the average difference between observed and predicted phenotypes. The results showed that the modification of the original boosting algorithm could be run in 1% of the time used with the original algorithm and with negligible differences in accuracy

  20. On factoring RSA modulus using random-restart hill-climbing algorithm and Pollard’s rho algorithm

    Science.gov (United States)

    Budiman, M. A.; Rachmawati, D.

    2017-12-01

    The security of the widely-used RSA public key cryptography algorithm depends on the difficulty of factoring a big integer into two large prime numbers. For many years, the integer factorization problem has been intensively and extensively studied in the field of number theory. As a result, a lot of deterministic algorithms such as Euler’s algorithm, Kraitchik’s, and variants of Pollard’s algorithms have been researched comprehensively. Our study takes a rather uncommon approach: rather than making use of intensive number theories, we attempt to factorize RSA modulus n by using random-restart hill-climbing algorithm, which belongs the class of metaheuristic algorithms. The factorization time of RSA moduli with different lengths is recorded and compared with the factorization time of Pollard’s rho algorithm, which is a deterministic algorithm. Our experimental results indicates that while random-restart hill-climbing algorithm is an acceptable candidate to factorize smaller RSA moduli, the factorization speed is much slower than that of Pollard’s rho algorithm.

  1. Antibiotic Residues - A Global Health Hazard

    Directory of Open Access Journals (Sweden)

    Nisha A.R.

    Full Text Available Use of Antibiotic that might result in deposition of residues in meat, milk and eggs must not be permitted in food intended for human consumption. If use of antibiotics is necessary as in prevention and treatment of animal diseases, a withholding period must be observed until the residues are negligible or no longer detected. The use of antibiotics to bring about improved performance in growth and feed efficiency, to synchronize or control of reproductive cycle and breeding performance also often lead to harmful residual effects. Concern over antibiotic residues in food of animal origin occurs in two times; one which produces potential threat to direct toxicity in human, second is whether the low levels of antibiotic exposure would result in alteration of microflora, cause disease and the possible development of resistant strains which cause failure of antibiotic therapy in clinical situations. A withdrawal period is established to safeguard human from exposure of antibiotic added food. The withdrawal time is the time required for the residue of toxicological concern to reach safe concentration as defined by tolerance. It is the interval from the time an animal is removed from medication until permitted time of slaughter. Heavy responsibility is placed on the veterinarian and livestock producer to observe the period for a withdrawal of a drug prior to slaughter to assure that illegal concentration of drug residue in meat, milk and egg do not occur. Use of food additives may improve feed efficiency 17% in beef cattle, 10% in lambs, 15% in poultry and 15% in swine. But their indiscriminate use will produce toxicity in consumers. WHO and FAO establish tolerances for a drug, pesticide or other chemical in the relevant tissues of food producing animals. The tolerance is the tissue concentration below, which a marker residue for the drug or chemical must fall in the target tissue before that animal edible tissues are considered safe for human

  2. Sustainable System for Residual Hazards Management

    International Nuclear Information System (INIS)

    Kevin M. Kostelnik; James H. Clarke; Jerry L. Harbour

    2004-01-01

    Hazardous, radioactive and other toxic substances have routinely been generated and subsequently disposed of in the shallow subsurface throughout the world. Many of today's waste management techniques do not eliminate the problem, but rather only concentrate or contain the hazardous contaminants. Residual hazards result from the presence of hazardous and/or contaminated material that remains on-site following active operations or the completion of remedial actions. Residual hazards pose continued risk to humans and the environment and represent a significant and chronic problem that require continuous long-term management (i.e. >1000 years). To protect human health and safeguard the natural environment, a sustainable system is required for the proper management of residual hazards. A sustainable system for the management of residual hazards will require the integration of engineered, institutional and land-use controls to isolate residual contaminants and thus minimize the associated hazards. Engineered controls are physical modifications to the natural setting and ecosystem, including the site, facility, and/or the residual materials themselves, in order to reduce or eliminate the potential for exposure to contaminants of concern (COCs). Institutional controls are processes, instruments, and mechanisms designed to influence human behavior and activity. System failure can involve hazardous material escaping from the confinement because of system degradation (i.e., chronic or acute degradation) or by external intrusion of the biosphere into the contaminated material because of the loss of institutional control. An ongoing analysis of contemporary and historic sites suggests that the significance of the loss of institutional controls is a critical pathway because decisions made during the operations/remedial action phase, as well as decisions made throughout the residual hazards management period, are key to the long-term success of the prescribed system. In fact

  3. 40 CFR 180.511 - Buprofezin; tolerances for residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Buprofezin; tolerances for residues... § 180.511 Buprofezin; tolerances for residues. (a) General. Tolerances are established for residues of buprofezin, including its metabolites and degradates in or on the commodities in the table below. Compliance...

  4. Opposition-Based Adaptive Fireworks Algorithm

    Directory of Open Access Journals (Sweden)

    Chibing Gong

    2016-07-01

    Full Text Available A fireworks algorithm (FWA is a recent swarm intelligence algorithm that is inspired by observing fireworks explosions. An adaptive fireworks algorithm (AFWA proposes additional adaptive amplitudes to improve the performance of the enhanced fireworks algorithm (EFWA. The purpose of this paper is to add opposition-based learning (OBL to AFWA with the goal of further boosting performance and achieving global optimization. Twelve benchmark functions are tested in use of an opposition-based adaptive fireworks algorithm (OAFWA. The final results conclude that OAFWA significantly outperformed EFWA and AFWA in terms of solution accuracy. Additionally, OAFWA was compared with a bat algorithm (BA, differential evolution (DE, self-adapting control parameters in differential evolution (jDE, a firefly algorithm (FA, and a standard particle swarm optimization 2011 (SPSO2011 algorithm. The research results indicate that OAFWA ranks the highest of the six algorithms for both solution accuracy and runtime cost.

  5. Managing woodwaste: Yield from residue

    Energy Technology Data Exchange (ETDEWEB)

    Nielson, E. [LNS Services, Inc., North Vancouver, British Columbia (Canada); Rayner, S. [Pacific Waste Energy Inc., Burnaby, British Columbia (Canada)

    1993-12-31

    Historically, the majority of sawmill waste has been burned or buried for the sole purpose of disposal. In most jurisdictions, environmental legislation will prohibit, or render uneconomic, these practices. Many reports have been prepared to describe the forest industry`s residue and its environmental effect; although these help those looking for industry-wide or regional solutions, such as electricity generation, they have limited value for the mill manager, who has the on-hands responsibility for generation and disposal of the waste. If the mill manager can evaluate waste streams and break them down into their usable components, he can find niche market solutions for portions of the plant residue and redirect waste to poor/no-return, rather than disposal-cost, end uses. In the modern mill, residue is collected at the individual machine centre by waste conveyors that combine and mix sawdust, shavings, bark, etc. and send the result to the hog-fuel pile. The mill waste system should be analyzed to determine the measures that can improve the quality of residues and determine the volumes of any particular category before the mixing, mentioned above, occurs. After this analysis, the mill may find a niche market for a portion of its woodwaste.

  6. Leptogenesis and residual CP symmetry

    International Nuclear Information System (INIS)

    Chen, Peng; Ding, Gui-Jun; King, Stephen F.

    2016-01-01

    We discuss flavour dependent leptogenesis in the framework of lepton flavour models based on discrete flavour and CP symmetries applied to the type-I seesaw model. Working in the flavour basis, we analyse the case of two general residual CP symmetries in the neutrino sector, which corresponds to all possible semi-direct models based on a preserved Z 2 in the neutrino sector, together with a CP symmetry, which constrains the PMNS matrix up to a single free parameter which may be fixed by the reactor angle. We systematically study and classify this case for all possible residual CP symmetries, and show that the R-matrix is tightly constrained up to a single free parameter, with only certain forms being consistent with successful leptogenesis, leading to possible connections between leptogenesis and PMNS parameters. The formalism is completely general in the sense that the two residual CP symmetries could result from any high energy discrete flavour theory which respects any CP symmetry. As a simple example, we apply the formalism to a high energy S 4 flavour symmetry with a generalized CP symmetry, broken to two residual CP symmetries in the neutrino sector, recovering familiar results for PMNS predictions, together with new results for flavour dependent leptogenesis.

  7. Opposite Degree Algorithm and Its Applications

    Directory of Open Access Journals (Sweden)

    Xiao-Guang Yue

    2015-12-01

    Full Text Available The opposite (Opposite Degree, referred to as OD algorithm is an intelligent algorithm proposed by Yue Xiaoguang et al. Opposite degree algorithm is mainly based on the concept of opposite degree, combined with the idea of design of neural network and genetic algorithm and clustering analysis algorithm. The OD algorithm is divided into two sub algorithms, namely: opposite degree - numerical computation (OD-NC algorithm and opposite degree - Classification computation (OD-CC algorithm.

  8. Actinide recovery from pyrochemical residues

    International Nuclear Information System (INIS)

    Avens, L.R.; Clifton, D.G.; Vigil, A.R.

    1984-01-01

    A new process for recovery of plutonium and americium from pyrochemical waste has been demonstrated. It is based on chloride solution anion exchange at low acidity, which eliminates corrosive HCl fumes. Developmental experiments of the process flowsheet concentrated on molten salt extraction (MSE) residues and gave >95% plutonium and >90% americium recovery. The recovered plutonium contained 6 = from high chloride-low acid solution. Americium and other metals are washed from the ion exchange column with 1N HNO 3 -4.8M NaCl. The plutonium is recovered, after elution, via hydroxide precipitation, while the americium is recovered via NaHCO 3 precipitation. All filtrates from the process are discardable as low-level contaminated waste. Production-scale experiments are now in progress for MSE residues. Flow sheets for actinide recovery from electrorefining and direct oxide reduction residues are presented and discussed

  9. Actinide recovery from pyrochemical residues

    International Nuclear Information System (INIS)

    Avens, L.R.; Clifton, D.G.; Vigil, A.R.

    1985-05-01

    We demonstrated a new process for recovering plutonium and americium from pyrochemical waste. The method is based on chloride solution anion exchange at low acidity, or acidity that eliminates corrosive HCl fumes. Developmental experiments of the process flow chart concentrated on molten salt extraction (MSE) residues and gave >95% plutonium and >90% americium recovery. The recovered plutonium contained 6 2- from high-chloride low-acid solution. Americium and other metals are washed from the ion exchange column with lN HNO 3 -4.8M NaCl. After elution, plutonium is recovered by hydroxide precipitation, and americium is recovered by NaHCO 3 precipitation. All filtrates from the process can be discardable as low-level contaminated waste. Production-scale experiments are in progress for MSE residues. Flow charts for actinide recovery from electro-refining and direct oxide reduction residues are presented and discussed

  10. 40 CFR 180.342 - Chlorpyrifos; tolerances for residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Chlorpyrifos; tolerances for residues... § 180.342 Chlorpyrifos; tolerances for residues. (a) General. (1) Tolerances are established for residues of the pesticide chlorpyrifos per se (O,O-diethyl- O-(3,5,6-trichloro-2-pyridyl) phosphorothioate...

  11. Trends and advances in pesticide residue analysis | Yeboah ...

    African Journals Online (AJOL)

    The nature, origin and the economic significance of pesticide residues are reviewed to underscore the need for countries to develop the ability and capacity to monitor pesticide residues. An overview of pesticide residues analytical procedures is also presented with emphasis on thin layer chromatography (TLC) as an ...

  12. 40 CFR 180.106 - Diuron; tolerances for residues.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 23 2010-07-01 2010-07-01 false Diuron; tolerances for residues. 180... Diuron; tolerances for residues. (a) General. Tolerances are established for the combined residues of the herbicide diuron, 3-(3,4-dichlorophenyl)-1,1-dimethylurea and its metabolites convertible to 3,4...

  13. Laboratory studies of the properties of in-situ burn residues: chemical composition of residues

    International Nuclear Information System (INIS)

    Trudel, B.K.; Buist, I.A.; Schatzke, D.; Aurand, D.

    1996-01-01

    The chemical composition of the residue from small-scale burns of thick oil slicks was studied. The objective was to describe the changes in chemical composition in oils burning on water and to determine how these changes were influenced by the condition of the burn. Small-scale test burns involved burning 40-cm diameter pools of oil on water. A range of eight oil types including seven crude oils and an automotive diesel were burned. For each oil, slicks of fresh oil of three different thicknesses were tested. Two of the oils were tested before and after weathering. Results showed that the composition of the residue differed greatly from the parent oil. Asphaltenes, high-boiling-point aromatics and resins remained concentrated in the burn residue. The burning of slicks appeared to remove most of the lower-molecular weight aromatic hydrocarbons which included the more toxic and more bioavailable components of the crude oils. 11 refs., 6 tabs

  14. An evidence-based treatment algorithm for colorectal polyp cancers: results from the Scottish Screen-detected Polyp Cancer Study (SSPoCS).

    Science.gov (United States)

    Richards, C H; Ventham, N T; Mansouri, D; Wilson, M; Ramsay, G; Mackay, C D; Parnaby, C N; Smith, D; On, J; Speake, D; McFarlane, G; Neo, Y N; Aitken, E; Forrest, C; Knight, K; McKay, A; Nair, H; Mulholland, C; Robertson, J H; Carey, F A; Steele, Rjc

    2018-02-01

    Colorectal polyp cancers present clinicians with a treatment dilemma. Decisions regarding whether to offer segmental resection or endoscopic surveillance are often taken without reference to good quality evidence. The aim of this study was to develop a treatment algorithm for patients with screen-detected polyp cancers. This national cohort study included all patients with a polyp cancer identified through the Scottish Bowel Screening Programme between 2000 and 2012. Multivariate regression analysis was used to assess the impact of clinical, endoscopic and pathological variables on the rate of adverse events (residual tumour in patients undergoing segmental resection or cancer-related death or disease recurrence in any patient). These data were used to develop a clinically relevant treatment algorithm. 485 patients with polyp cancers were included. 186/485 (38%) underwent segmental resection and residual tumour was identified in 41/186 (22%). The only factor associated with an increased risk of residual tumour in the bowel wall was incomplete excision of the original polyp (OR 5.61, p=0.001), while only lymphovascular invasion was associated with an increased risk of lymph node metastases (OR 5.95, p=0.002). When patients undergoing segmental resection or endoscopic surveillance were considered together, the risk of adverse events was significantly higher in patients with incomplete excision (OR 10.23, p<0.001) or lymphovascular invasion (OR 2.65, p=0.023). A policy of surveillance is adequate for the majority of patients with screen-detected colorectal polyp cancers. Consideration of segmental resection should be reserved for those with incomplete excision or evidence of lymphovascular invasion. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.

  15. Analysis and radiological assessment of residues containing NORM materials resulting from earlier activities including modelling of typical industrial residues. Pt. 1. Historical investigation of the radiological relevance of NORM residues and concepts for site identification

    International Nuclear Information System (INIS)

    Reichelt, Andreas; Niedermayer, Matthias; Sitte, Beate; Hamel, Peter Michael

    2007-01-01

    Natural radionuclides are part of the human environment and of the raw materials used. Technical processes may cause their accumulation in residues, and the result will be so-called NORM materials (Naturally occurring radioactive material). The amended Radiation Protection Ordinance (StrlSchV 2001) specifies how the public should be protected, but there are also residues dating back before the issuing of the StrlSchV 2001, the so-called NORM residues. The project intended to assess the risks resulting from these residues. It comprises four parts. Part 1 was for clarification of the radiological relevance of NORM residues and for the development of concepts to detect them. The criterion for their radiological relevance was their activity per mass unit and the material volume accumulated through the centuries. The former was calculated from a wide bibliographic search in the relevant literature on radiation protection, while the mass volume was obtained by a detailed historical search of the consumption of materials that may leave NORM residues. These are, in particular, residues from coal and ore mining and processing. To identify concrete sites, relevant data sources were identified, and a concept for identification of concrete NORM residues was developed on this basis. (orig.) [de

  16. Multi-Level iterative methods in computational plasma physics

    International Nuclear Information System (INIS)

    Knoll, D.A.; Barnes, D.C.; Brackbill, J.U.; Chacon, L.; Lapenta, G.

    1999-01-01

    Plasma physics phenomena occur on a wide range of spatial scales and on a wide range of time scales. When attempting to model plasma physics problems numerically the authors are inevitably faced with the need for both fine spatial resolution (fine grids) and implicit time integration methods. Fine grids can tax the efficiency of iterative methods and large time steps can challenge the robustness of iterative methods. To meet these challenges they are developing a hybrid approach where multigrid methods are used as preconditioners to Krylov subspace based iterative methods such as conjugate gradients or GMRES. For nonlinear problems they apply multigrid preconditioning to a matrix-few Newton-GMRES method. Results are presented for application of these multilevel iterative methods to the field solves in implicit moment method PIC, multidimensional nonlinear Fokker-Planck problems, and their initial efforts in particle MHD

  17. Evaluation of gunshot residue (GSR) evidence: Surveys of prevalence of GSR on clothing and frequency of residue types.

    Science.gov (United States)

    Hannigan, Thomas J; McDermott, Sean D; Greaney, Claire M; O'Shaughnessy, John; O'Brien, Cliona M

    2015-12-01

    The evaluative approach is a logical approach to interpreting scientific findings in criminal cases, applying knowledge regarding the transfer, persistence and recovery of particulate material. The application of this approach to interpreting the finding of gunshot residue on the clothing of a suspect requires knowledge of background levels of GSR on clothing and on the frequency of different residue types in a particular environment. The cuffs of 100 upper outer garments submitted to a forensic laboratory in connection with non-firearms offences were sampled for gunshot residue. No 3-component lead/antimony/barium particles were found on 98 of them. Two 3-component particles were found on one of them and one 3-component particle was found on another. The frequency of occurrence of various particle types regarded as consistent with GSR was also explored. The findings show that, while 3-component particles were somewhat more likely to be encountered by chance on clothing than on hands, they are still relatively uncommon events. To investigate the frequency of occurrence of particular residue types, 100 discharged rounds of ammunition recovered at crime scenes were sampled and the types of residue present were determined. The results show that some residue types are significantly more common than others. Both sets of data will be of value in evaluating the significance of finding GSR on clothing of suspects in criminal cases. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  18. Algorithmic phase diagrams

    Science.gov (United States)

    Hockney, Roger

    1987-01-01

    Algorithmic phase diagrams are a neat and compact representation of the results of comparing the execution time of several algorithms for the solution of the same problem. As an example, the recent results are shown of Gannon and Van Rosendale on the solution of multiple tridiagonal systems of equations in the form of such diagrams. The act of preparing these diagrams has revealed an unexpectedly complex relationship between the best algorithm and the number and size of the tridiagonal systems, which was not evident from the algebraic formulae in the original paper. Even so, for a particular computer, one diagram suffices to predict the best algorithm for all problems that are likely to be encountered the prediction being read directly from the diagram without complex calculation.

  19. Recursive forgetting algorithms

    DEFF Research Database (Denmark)

    Parkum, Jens; Poulsen, Niels Kjølstad; Holst, Jan

    1992-01-01

    In the first part of the paper, a general forgetting algorithm is formulated and analysed. It contains most existing forgetting schemes as special cases. Conditions are given ensuring that the basic convergence properties will hold. In the second part of the paper, the results are applied...... to a specific algorithm with selective forgetting. Here, the forgetting is non-uniform in time and space. The theoretical analysis is supported by a simulation example demonstrating the practical performance of this algorithm...

  20. Disposal of Rocky Flats residues as waste

    International Nuclear Information System (INIS)

    Dustin, D.F.; Sendelweck, V.S.

    1993-01-01

    Work is underway at the Rocky Flats Plant to evaluate alternatives for the removal of a large inventory of plutonium-contaminated residues from the plant. One alternative under consideration is to package the residues as transuranic wastes for ultimate shipment to the Waste Isolation Pilot Plant. Current waste acceptance criteria and transportation regulations require that approximately 1000 cubic yards of residues be repackaged to produce over 20,000 cubic yards of WIPP certified waste. The major regulatory drivers leading to this increase in waste volume are the fissile gram equivalent, surface radiation dose rate, and thermal power limits. In the interest of waste minimization, analyses have been conducted to determine, for each residue type, the controlling criterion leading to the volume increase, the impact of relaxing that criterion on subsequent waste volume, and the means by which rules changes may be implemented. The results of this study have identified the most appropriate changes to be proposed in regulatory requirements in order to minimize the costs of disposing of Rocky Flats residues as transuranic wastes

  1. Multi-objective optimization of in-situ bioremediation of groundwater using a hybrid metaheuristic technique based on differential evolution, genetic algorithms and simulated annealing

    Directory of Open Access Journals (Sweden)

    Kumar Deepak

    2015-12-01

    Full Text Available Groundwater contamination due to leakage of gasoline is one of the several causes which affect the groundwater environment by polluting it. In the past few years, In-situ bioremediation has attracted researchers because of its ability to remediate the contaminant at its site with low cost of remediation. This paper proposed the use of a new hybrid algorithm to optimize a multi-objective function which includes the cost of remediation as the first objective and residual contaminant at the end of the remediation period as the second objective. The hybrid algorithm was formed by combining the methods of Differential Evolution, Genetic Algorithms and Simulated Annealing. Support Vector Machines (SVM was used as a virtual simulator for biodegradation of contaminants in the groundwater flow. The results obtained from the hybrid algorithm were compared with Differential Evolution (DE, Non Dominated Sorting Genetic Algorithm (NSGA II and Simulated Annealing (SA. It was found that the proposed hybrid algorithm was capable of providing the best solution. Fuzzy logic was used to find the best compromising solution and finally a pumping rate strategy for groundwater remediation was presented for the best compromising solution. The results show that the cost incurred for the best compromising solution is intermediate between the highest and lowest cost incurred for other non-dominated solutions.

  2. Residues from waste incineration. Final report. Rev. ed.

    Energy Technology Data Exchange (ETDEWEB)

    Astrup, T.; Juul Pedersen, A.; Hyks, J.; Frandsen, F.J.

    2010-04-15

    The overall objective of the project was to improve the understanding of the formation and characteristics of residues from waste incineration. This was done focusing on the importance of the waste input and the operational conditions of the furnace. Data and results obtained from the project have been discussed in this report according to the following three overall parts: i) mass flows and element distribution, ii) flue gas/particle partitioning and corrosion/deposition aspects, and iii) residue leaching. This has been done with the intent of structuring the discussion while tacitly acknowledging that these aspects are interrelated and cannot be separated. Overall, it was found that the waste input composition had significant impact of the characteristics of the generated residues. A similar correlation between operational conditions and residue characteristics could not be observed. Consequently, the project recommend that optimization of residue quality should focus on controlling the waste input composition. The project results showed that including specific waste materials (and thereby also excluding the same materials) may have significant effects on the residue composition, residue leaching, aerosol and deposit formation.It is specifically recommended to minimize Cl in the input waste. Based on the project results, it was found that a significant potential for optimization of waste incineration exist. (Author)

  3. Carbaryl residues in cottonseed products in the Philippines

    International Nuclear Information System (INIS)

    Pablo, F.E.

    1981-01-01

    Residues of carbaryl in cottonseed products were determined by spectrophotometry. Carbaryl residues in oil and cake were 0.83 and 0.04 mg/kg respectively. Parallel experiments with 14 C-carbaryl gave comparable results. 14 C-carbaryl residues in oil and cake were 0.42 and 0.15 mg/kg respectively. (author)

  4. Identification of mannose interacting residues using local composition.

    Directory of Open Access Journals (Sweden)

    Sandhya Agarwal

    Full Text Available BACKGROUND: Mannose binding proteins (MBPs play a vital role in several biological functions such as defense mechanisms. These proteins bind to mannose on the surface of a wide range of pathogens and help in eliminating these pathogens from our body. Thus, it is important to identify mannose interacting residues (MIRs in order to understand mechanism of recognition of pathogens by MBPs. RESULTS: This paper describes modules developed for predicting MIRs in a protein. Support vector machine (SVM based models have been developed on 120 mannose binding protein chains, where no two chains have more than 25% sequence similarity. SVM models were developed on two types of datasets: 1 main dataset consists of 1029 mannose interacting and 1029 non-interacting residues, 2 realistic dataset consists of 1029 mannose interacting and 10320 non-interacting residues. In this study, firstly, we developed standard modules using binary and PSSM profile of patterns and got maximum MCC around 0.32. Secondly, we developed SVM modules using composition profile of patterns and achieved maximum MCC around 0.74 with accuracy 86.64% on main dataset. Thirdly, we developed a model on a realistic dataset and achieved maximum MCC of 0.62 with accuracy 93.08%. Based on this study, a standalone program and web server have been developed for predicting mannose interacting residues in proteins (http://www.imtech.res.in/raghava/premier/. CONCLUSIONS: Compositional analysis of mannose interacting and non-interacting residues shows that certain types of residues are preferred in mannose interaction. It was also observed that residues around mannose interacting residues have a preference for certain types of residues. Composition of patterns/peptide/segment has been used for predicting MIRs and achieved reasonable high accuracy. It is possible that this novel strategy may be effective to predict other types of interacting residues. This study will be useful in annotating the function

  5. Agency and Algorithms

    Directory of Open Access Journals (Sweden)

    Hanns Holger Rutz

    2016-11-01

    Full Text Available Although the concept of algorithms has been established a long time ago, their current topicality indicates a shift in the discourse. Classical definitions based on logic seem to be inadequate to describe their aesthetic capabilities. New approaches stress their involvement in material practices as well as their incompleteness. Algorithmic aesthetics can no longer be tied to the static analysis of programs, but must take into account the dynamic and experimental nature of coding practices. It is suggested that the aesthetic objects thus produced articulate something that could be called algorithmicity or the space of algorithmic agency. This is the space or the medium – following Luhmann’s form/medium distinction – where human and machine undergo mutual incursions. In the resulting coupled “extimate” writing process, human initiative and algorithmic speculation cannot be clearly divided out any longer. An observation is attempted of defining aspects of such a medium by drawing a trajectory across a number of sound pieces. The operation of exchange between form and medium I call reconfiguration and it is indicated by this trajectory. 

  6. Residual stress relaxation measurements across interfaces at macro-and micro-scales using slitting and DIC

    Energy Technology Data Exchange (ETDEWEB)

    Blair, A; Daynes, N; Hamilton, D; Horne, G; Hodgson, D Z L; Shterenlikht, A [Department of Mechanical Engineering, University of Bristol, Bristol BS8 1TR (United Kingdom); Heard, P J; Scott, T B, E-mail: mexas@bristol.ac.u [Interface Analysis Centre, University of Bristol, Bristol BS2 8BS (United Kingdom)

    2009-08-01

    In this paper digital image correlation is used to measure relaxation of residual stresses across an interface. On the macro scale the method is applied to a tri-layer bonded aluminium sample, where the middle layer is in tension and the top and the bottom layers are in compression. High contrast speckle pattern was sprayed onto the surface. The relaxation was done with the slitting saw. Three dimensional image correlation was used. On the micro scale the technique was applied to a heat treated large grain brass loaded in tension. Mechanical and electro polishing was used for surface preparation. A focused ion beam was used for slitting across a grain boundary and for imaging. Grain orientation was measured using electron back-scattering diffraction. Two dimensional image correlation was employed. In all macro- and micro-scale experiments the range of measured relaxation was sub-pixel, almost at the limit of the resolution of the image correlation algorithms. In the macro-scale experiments, the limiting factor was low residual stress, due to low shear strength of the Araldite glue used for bonding. Finite element simulation of the relaxation agreed only qualitatively with the experimental results at both size scales. The methodology is intended for use with inverse methods, i.e. the measured relaxation is applied as the boundary conditions to an appropriate FE model which produces stresses equal to the relaxed residual stresses, but with opposite sign. The main conclusion is that the digital image correlation method could be used to measure relaxation caused by slitting in heterogeneous materials and structures at both macro- and micro-scales. However, the repeatability of the techniques needs to be improved before residual stresses can be determined confidently. Acknowledgments The authors gratefully acknowledge Airbus UK for provision of materials. They thank Dr Richard Burguete, Airbus UK, and Prof Peter Flewitt, Department of Physics, University of Bristol, for

  7. Residual stress relaxation measurements across interfaces at macro-and micro-scales using slitting and DIC

    International Nuclear Information System (INIS)

    Blair, A; Daynes, N; Hamilton, D; Horne, G; Hodgson, D Z L; Shterenlikht, A; Heard, P J; Scott, T B

    2009-01-01

    In this paper digital image correlation is used to measure relaxation of residual stresses across an interface. On the macro scale the method is applied to a tri-layer bonded aluminium sample, where the middle layer is in tension and the top and the bottom layers are in compression. High contrast speckle pattern was sprayed onto the surface. The relaxation was done with the slitting saw. Three dimensional image correlation was used. On the micro scale the technique was applied to a heat treated large grain brass loaded in tension. Mechanical and electro polishing was used for surface preparation. A focused ion beam was used for slitting across a grain boundary and for imaging. Grain orientation was measured using electron back-scattering diffraction. Two dimensional image correlation was employed. In all macro- and micro-scale experiments the range of measured relaxation was sub-pixel, almost at the limit of the resolution of the image correlation algorithms. In the macro-scale experiments, the limiting factor was low residual stress, due to low shear strength of the Araldite glue used for bonding. Finite element simulation of the relaxation agreed only qualitatively with the experimental results at both size scales. The methodology is intended for use with inverse methods, i.e. the measured relaxation is applied as the boundary conditions to an appropriate FE model which produces stresses equal to the relaxed residual stresses, but with opposite sign. The main conclusion is that the digital image correlation method could be used to measure relaxation caused by slitting in heterogeneous materials and structures at both macro- and micro-scales. However, the repeatability of the techniques needs to be improved before residual stresses can be determined confidently. Acknowledgments The authors gratefully acknowledge Airbus UK for provision of materials. They thank Dr Richard Burguete, Airbus UK, and Prof Peter Flewitt, Department of Physics, University of Bristol, for

  8. Residual stress measurement of large scaled welded pipe using neutron diffraction method. Effect of SCC crack propagation and repair weld on residual stress distribution

    International Nuclear Information System (INIS)

    Suzuki, Hiroshi; Katsuyama, Jinya; Tobita, Tohru; Morii, Yukio

    2011-01-01

    The RESA-1 neutron engineering diffractometer in the JRR-3 (Japan Research Reactor No.3) at the Japan Atomic Energy Agency, which is used for stress measurements, was upgraded to realize residual stress measurements of large scaled mechanical components. A series of residual stress measurements was made to obtain through-thickness residual stress distributions in a Type 304 stainless steel butt-welded pipe of 500A-sch.80 using the upgraded RESA-1 diffractometer. We evaluated effects of crack propagation such as stress corrosion cracking (SCC) and a part-circumference repair weld on the residual stress distributions induced by girth welding. Measured residual stress distributions near original girth weld revealed good agreement with typical results shown in some previous works using finite element method, deep hole drilling as well as neutron diffraction. After introducing a mock crack with 10 mm depth in the heat affected zone on the inside wall of the pipe by electro discharge machining, the axial residual stresses were found to be released in the part of the mock crack. However, changes in the through-wall bending stress component and the self-equilibrated stress component were negligible and hence the axial residual stress distribution in the ligament was remained in the original residual stresses near girth weld without the mock crack. Furthermore, changes in hoop and radial residual stress were also small. The residual stress distributions after a part repair welding on the outer circumference of the girth weld were significantly different from residual stress distributions near the original girth weld. The through-thickness average axial residual stress was increased due to increase of the tensile membrane stress and mitigation of the bending stress after repair welding. Throughout above studies, we evidenced that the neutron diffraction technique is useful and powerful tool for measuring residual stress distributions in large as well as thick mechanical

  9. Adsorption Property and Mechanism of Oxytetracycline onto Willow Residues

    Directory of Open Access Journals (Sweden)

    Di Wang

    2017-12-01

    Full Text Available To elucidate the adsorption property and the mechanism of plant residues to reduce oxytetracycline (OTC, the adsorption of OTC onto raw willow roots (WR-R, stems (WS-R, leaves (WL-R, and adsorption onto desugared willow roots (WR-D, stems (WS-D, and leaves (WL-D were investigated. The structural characterization was analyzed by scanning electron microscopy, Fourier-transform infrared spectra, and an elemental analyzer. OTC adsorption onto the different tissues of willow residues was compared and correlated with their structures. The adsorption kinetics of OTC onto willow residues was found to follow the pseudo-first-order model. The isothermal adsorption process of OTC onto the different tissues of willow residues followed the Langmuir and Freundlich model and the process was also a spontaneous endothermic reaction, which was mainly physical adsorption. After the willow residues were desugared, the polarity decreased and the aromaticity increased, which explained why the adsorption amounts of the desugared willow residues were higher than those of the unmodified residues. These observations suggest that the raw and modified willow residues have great potential as adsorbents to remove organic pollutants.

  10. Electroremediation of air pollution control residues in a continuous reactor

    DEFF Research Database (Denmark)

    Jensen, Pernille Erland; Ferreira, Célia M. D.; Hansen, Henrik K.

    2010-01-01

    Air pollution control (APC) residue from municipal solid waste incineration is considered hazardous waste due to its alkalinity and high content of salts and mobile heavy metals. Various solutions for the handling of APC-residue exist, however most commercial solutions involve landfilling. A demand...... were made with raw residue, water-washed residue, acid washed residue and acid-treated residue with emphasis on reduction of heavy metal mobility. Main results indicate that the reactor successfully removes toxic elements lead, copper, cadmium and zinc from the feed stream, suggesting...

  11. Real-Coded Quantum-Inspired Genetic Algorithm-Based BP Neural Network Algorithm

    Directory of Open Access Journals (Sweden)

    Jianyong Liu

    2015-01-01

    Full Text Available The method that the real-coded quantum-inspired genetic algorithm (RQGA used to optimize the weights and threshold of BP neural network is proposed to overcome the defect that the gradient descent method makes the algorithm easily fall into local optimal value in the learning process. Quantum genetic algorithm (QGA is with good directional global optimization ability, but the conventional QGA is based on binary coding; the speed of calculation is reduced by the coding and decoding processes. So, RQGA is introduced to explore the search space, and the improved varied learning rate is adopted to train the BP neural network. Simulation test shows that the proposed algorithm is effective to rapidly converge to the solution conformed to constraint conditions.

  12. Soil water evaporation and crop residues

    Science.gov (United States)

    Crop residues have value when left in the field and also when removed from the field and sold as a commodity. Reducing soil water evaporation (E) is one of the benefits of leaving crop residues in place. E was measured beneath a corn canopy at the soil suface with nearly full coverage by corn stover...

  13. Cement production from coal conversion residues

    International Nuclear Information System (INIS)

    Brown, L.D.; Clavenna, L.R.; Eakman, J.M.; Nahas, N.C.

    1981-01-01

    Cement is produced by feeding residue solids containing carbonaceous material and ash constituents obtained from converting a carbonaceous feed material into liquids and/or gases into a cement-making zone and burning the carbon in the residue solids to supply at least a portion of the energy required to convert the solids into cement

  14. A cluster algorithm for graphs

    NARCIS (Netherlands)

    S. van Dongen

    2000-01-01

    textabstractA cluster algorithm for graphs called the emph{Markov Cluster algorithm (MCL~algorithm) is introduced. The algorithm provides basically an interface to an algebraic process defined on stochastic matrices, called the MCL~process. The graphs may be both weighted (with nonnegative weight)

  15. Residual stresses in Inconel 718 engine disks

    Directory of Open Access Journals (Sweden)

    Dahan Yoann

    2014-01-01

    Full Text Available Aubert&Duval has developed a methodology to establish a residual stress model for Inconel 718 engine discs. To validate the thermal, mechanical and metallurgical parts of the model, trials on lab specimens with specific geometry were carried out. These trials allow a better understanding of the residual stress distribution and evolution during different processes (quenching, ageing, machining. A comparison between experimental and numerical results reveals the residual stresses model accuracy. Aubert&Duval has also developed a mechanical properties prediction model. Coupled with the residual stress prediction model, Aubert&Duval can now propose improvements to the process of manufacturing in Inconel 718 engine disks. This model enables Aubert&Duval customers and subcontractors to anticipate distortions issues during machining. It could also be usedt to optimise the engine disk life.

  16. Improving residue-residue contact prediction via low-rank and sparse decomposition of residue correlation matrix.

    Science.gov (United States)

    Zhang, Haicang; Gao, Yujuan; Deng, Minghua; Wang, Chao; Zhu, Jianwei; Li, Shuai Cheng; Zheng, Wei-Mou; Bu, Dongbo

    2016-03-25

    Strategies for correlation analysis in protein contact prediction often encounter two challenges, namely, the indirect coupling among residues, and the background correlations mainly caused by phylogenetic biases. While various studies have been conducted on how to disentangle indirect coupling, the removal of background correlations still remains unresolved. Here, we present an approach for removing background correlations via low-rank and sparse decomposition (LRS) of a residue correlation matrix. The correlation matrix can be constructed using either local inference strategies (e.g., mutual information, or MI) or global inference strategies (e.g., direct coupling analysis, or DCA). In our approach, a correlation matrix was decomposed into two components, i.e., a low-rank component representing background correlations, and a sparse component representing true correlations. Finally the residue contacts were inferred from the sparse component of correlation matrix. We trained our LRS-based method on the PSICOV dataset, and tested it on both GREMLIN and CASP11 datasets. Our experimental results suggested that LRS significantly improves the contact prediction precision. For example, when equipped with the LRS technique, the prediction precision of MI and mfDCA increased from 0.25 to 0.67 and from 0.58 to 0.70, respectively (Top L/10 predicted contacts, sequence separation: 5 AA, dataset: GREMLIN). In addition, our LRS technique also consistently outperforms the popular denoising technique APC (average product correction), on both local (MI_LRS: 0.67 vs MI_APC: 0.34) and global measures (mfDCA_LRS: 0.70 vs mfDCA_APC: 0.67). Interestingly, we found out that when equipped with our LRS technique, local inference strategies performed in a comparable manner to that of global inference strategies, implying that the application of LRS technique narrowed down the performance gap between local and global inference strategies. Overall, our LRS technique greatly facilitates

  17. Recent advances in residual stress measurement

    International Nuclear Information System (INIS)

    Withers, P.J.; Turski, M.; Edwards, L.; Bouchard, P.J.; Buttle, D.J.

    2008-01-01

    Until recently residual stresses have been included in structural integrity assessments of nuclear pressure vessels and piping in a very primitive manner due to the lack of reliable residual stress measurement or prediction tools. This situation is changing the capabilities of newly emerging destructive (i.e. the contour method) and non-destructive (i.e. magnetic and high-energy synchrotron X-ray strain mapping) residual stress measurement techniques for evaluating ferritic and austenitic pressure vessel components are contrasted against more well-established methods. These new approaches offer the potential for obtaining area maps of residual stress or strain in welded plants, mock-up components or generic test-pieces. The mapped field may be used directly in structural integrity calculations, or indirectly to validate finite element process/structural models on which safety cases for pressurised nuclear systems are founded. These measurement methods are complementary in terms of application to actual plant, cost effectiveness and measurements in thick sections. In each case an exemplar case study is used to illustrate the method and to highlight its particular capabilities

  18. Unsupervised learning algorithms

    CERN Document Server

    Aydin, Kemal

    2016-01-01

    This book summarizes the state-of-the-art in unsupervised learning. The contributors discuss how with the proliferation of massive amounts of unlabeled data, unsupervised learning algorithms, which can automatically discover interesting and useful patterns in such data, have gained popularity among researchers and practitioners. The authors outline how these algorithms have found numerous applications including pattern recognition, market basket analysis, web mining, social network analysis, information retrieval, recommender systems, market research, intrusion detection, and fraud detection. They present how the difficulty of developing theoretically sound approaches that are amenable to objective evaluation have resulted in the proposal of numerous unsupervised learning algorithms over the past half-century. The intended audience includes researchers and practitioners who are increasingly using unsupervised learning algorithms to analyze their data. Topics of interest include anomaly detection, clustering,...

  19. On the Structural Context and Identification of Enzyme Catalytic Residues

    Directory of Open Access Journals (Sweden)

    Yu-Tung Chien

    2013-01-01

    Full Text Available Enzymes play important roles in most of the biological processes. Although only a small fraction of residues are directly involved in catalytic reactions, these catalytic residues are the most crucial parts in enzymes. The study of the fundamental and unique features of catalytic residues benefits the understanding of enzyme functions and catalytic mechanisms. In this work, we analyze the structural context of catalytic residues based on theoretical and experimental structure flexibility. The results show that catalytic residues have distinct structural features and context. Their neighboring residues, whether sequence or structure neighbors within specific range, are usually structurally more rigid than those of noncatalytic residues. The structural context feature is combined with support vector machine to identify catalytic residues from enzyme structure. The prediction results are better or comparable to those of recent structure-based prediction methods.

  20. Generalized-ensemble molecular dynamics and Monte Carlo algorithms beyond the limit of the multicanonical algorithm

    International Nuclear Information System (INIS)

    Okumura, Hisashi

    2010-01-01

    I review two new generalized-ensemble algorithms for molecular dynamics and Monte Carlo simulations of biomolecules, that is, the multibaric–multithermal algorithm and the partial multicanonical algorithm. In the multibaric–multithermal algorithm, two-dimensional random walks not only in the potential-energy space but also in the volume space are realized. One can discuss the temperature dependence and pressure dependence of biomolecules with this algorithm. The partial multicanonical simulation samples a wide range of only an important part of potential energy, so that one can concentrate the effort to determine a multicanonical weight factor only on the important energy terms. This algorithm has higher sampling efficiency than the multicanonical and canonical algorithms. (review)

  1. Total sulfur determination in residues of crude oil distillation using FT-IR/ATR and variable selection methods

    Science.gov (United States)

    Müller, Aline Lima Hermes; Picoloto, Rochele Sogari; Mello, Paola de Azevedo; Ferrão, Marco Flores; dos Santos, Maria de Fátima Pereira; Guimarães, Regina Célia Lourenço; Müller, Edson Irineu; Flores, Erico Marlon Moraes

    2012-04-01

    Total sulfur concentration was determined in atmospheric residue (AR) and vacuum residue (VR) samples obtained from petroleum distillation process by Fourier transform infrared spectroscopy with attenuated total reflectance (FT-IR/ATR) in association with chemometric methods. Calibration and prediction set consisted of 40 and 20 samples, respectively. Calibration models were developed using two variable selection models: interval partial least squares (iPLS) and synergy interval partial least squares (siPLS). Different treatments and pre-processing steps were also evaluated for the development of models. The pre-treatment based on multiplicative scatter correction (MSC) and the mean centered data were selected for models construction. The use of siPLS as variable selection method provided a model with root mean square error of prediction (RMSEP) values significantly better than those obtained by PLS model using all variables. The best model was obtained using siPLS algorithm with spectra divided in 20 intervals and combinations of 3 intervals (911-824, 823-736 and 737-650 cm-1). This model produced a RMSECV of 400 mg kg-1 S and RMSEP of 420 mg kg-1 S, showing a correlation coefficient of 0.990.

  2. Prediction of residual stress for dissimilar metals welding at nuclear power plants using fuzzy neural network models

    International Nuclear Information System (INIS)

    Na, Man Gyun; Kim, Jin Weon; Lim, Dong Hyuk

    2007-01-01

    A fuzzy neural network model is presented to predict residual stress for dissimilar metal welding under various welding conditions. The fuzzy neural network model, which consists of a fuzzy inference system and a neuronal training system, is optimized by a hybrid learning method that combines a genetic algorithm to optimize the membership function parameters and a least squares method to solve the consequent parameters. The data of finite element analysis are divided into four data groups, which are split according to two end-section constraints and two prediction paths. Four fuzzy neural network models were therefore applied to the numerical data obtained from the finite element analysis for the two end-section constraints and the two prediction paths. The fuzzy neural network models were trained with the aid of a data set prepared for training (training data), optimized by means of an optimization data set and verified by means of a test data set that was different (independent) from the training data and the optimization data. The accuracy of fuzzy neural network models is known to be sufficiently accurate for use in an integrity evaluation by predicting the residual stress of dissimilar metal welding zones

  3. Theoretic derivation of directed acyclic subgraph algorithm and comparisons with message passing algorithm

    Science.gov (United States)

    Ha, Jeongmok; Jeong, Hong

    2016-07-01

    This study investigates the directed acyclic subgraph (DAS) algorithm, which is used to solve discrete labeling problems much more rapidly than other Markov-random-field-based inference methods but at a competitive accuracy. However, the mechanism by which the DAS algorithm simultaneously achieves competitive accuracy and fast execution speed, has not been elucidated by a theoretical derivation. We analyze the DAS algorithm by comparing it with a message passing algorithm. Graphical models, inference methods, and energy-minimization frameworks are compared between DAS and message passing algorithms. Moreover, the performances of DAS and other message passing methods [sum-product belief propagation (BP), max-product BP, and tree-reweighted message passing] are experimentally compared.

  4. Evaluation of Residual Stresses using Ring Core Method

    Directory of Open Access Journals (Sweden)

    Holý S.

    2010-06-01

    Full Text Available The method for measuring residual stresses using ring-core method is described. Basic relations are given for residual stress measurement along the specimen depth and simplified method is described for average residual stress estimation in the drilled layer for known principal stress directions. The estimation of calculated coefficients using FEM is described. Comparison of method sensitivity is made with hole-drilling method. The device for method application is described and an example of experiment is introduced. The accuracy of method is discussed. The influence of strain gauge rosette misalignment to the evaluated residual stresses is performed using FEM.

  5. Shadow algorithms data miner

    CERN Document Server

    Woo, Andrew

    2012-01-01

    Digital shadow generation continues to be an important aspect of visualization and visual effects in film, games, simulations, and scientific applications. This resource offers a thorough picture of the motivations, complexities, and categorized algorithms available to generate digital shadows. From general fundamentals to specific applications, it addresses shadow algorithms and how to manage huge data sets from a shadow perspective. The book also examines the use of shadow algorithms in industrial applications, in terms of what algorithms are used and what software is applicable.

  6. Diversity-Guided Evolutionary Algorithms

    DEFF Research Database (Denmark)

    Ursem, Rasmus Kjær

    2002-01-01

    Population diversity is undoubtably a key issue in the performance of evolutionary algorithms. A common hypothesis is that high diversity is important to avoid premature convergence and to escape local optima. Various diversity measures have been used to analyze algorithms, but so far few...... algorithms have used a measure to guide the search. The diversity-guided evolutionary algorithm (DGEA) uses the wellknown distance-to-average-point measure to alternate between phases of exploration (mutation) and phases of exploitation (recombination and selection). The DGEA showed remarkable results...

  7. Contact-impact algorithms on parallel computers

    International Nuclear Information System (INIS)

    Zhong Zhihua; Nilsson, Larsgunnar

    1994-01-01

    Contact-impact algorithms on parallel computers are discussed within the context of explicit finite element analysis. The algorithms concerned include a contact searching algorithm and an algorithm for contact force calculations. The contact searching algorithm is based on the territory concept of the general HITA algorithm. However, no distinction is made between different contact bodies, or between different contact surfaces. All contact segments from contact boundaries are taken as a single set. Hierarchy territories and contact territories are expanded. A three-dimensional bucket sort algorithm is used to sort contact nodes. The defence node algorithm is used in the calculation of contact forces. Both the contact searching algorithm and the defence node algorithm are implemented on the connection machine CM-200. The performance of the algorithms is examined under different circumstances, and numerical results are presented. ((orig.))

  8. A review on quantum search algorithms

    Science.gov (United States)

    Giri, Pulak Ranjan; Korepin, Vladimir E.

    2017-12-01

    The use of superposition of states in quantum computation, known as quantum parallelism, has significant advantage in terms of speed over the classical computation. It is evident from the early invented quantum algorithms such as Deutsch's algorithm, Deutsch-Jozsa algorithm and its variation as Bernstein-Vazirani algorithm, Simon algorithm, Shor's algorithms, etc. Quantum parallelism also significantly speeds up the database search algorithm, which is important in computer science because it comes as a subroutine in many important algorithms. Quantum database search of Grover achieves the task of finding the target element in an unsorted database in a time quadratically faster than the classical computer. We review Grover's quantum search algorithms for a singe and multiple target elements in a database. The partial search algorithm of Grover and Radhakrishnan and its optimization by Korepin called GRK algorithm are also discussed.

  9. Effect of soil-bound residues of malathion on microbial activities

    International Nuclear Information System (INIS)

    Hussain, A.; Iqbal, Z.; Asi, M.R.; Tahira, R.; Chudhary, J.A.

    2001-01-01

    The effect of soil-bound residues of malathion on CO/sub 2/ evolution, dehydrogenase activity and some nitrogen transformations in a loam soil was investigated under laboratory conditions. The soil samples containing bound residues arising from 10 mg g-1 of the applied malathion were mixed in equal quantity with fresh soil and compared with solvent extracted control soil without bound residues (extracted in the same way as soil containing bound residues). Another control comprising un extracted fresh soil without bound residues was also kept to study the effect of solvent extraction on the biological activity. Rate of Carbon mineralization (CO/sub 2/ evolution) was decreased in the presence of soil-bound residues of malathion. Bound residues also affected dehydrogenase activity of soil. Over 40% inhibition of dehydrogenase activity was observed after 4 days and the inhibition persisted at least for 12 days. Nitrogen mineralization was stimulated in soil containing bound residues of malathion and this stimulatory effect increased with time of incubation. Nitrification was partially inhibited in the presence of soil-bound residues of malathion. The inhibitory effect of the soil-bound residues on nitrification did not show much variation with time. The soil-bound residues did not affect denitrification rate (N/sub 2/O evolution). Nitrogen fixation (acetylene reduction) was partially inhibited in soil amended with bound residues of malathion and the inhibitory effect persisted for at least one week. In general, soil bound residues of malathion inhibited CO/sub 2/ evolution, dehydrogenase activity, nitrification and nitrogen fixation while mineralization of nitrogen was stimulated. Denitrification was not affected by the applied insecticide. (author)

  10. Magnet sorting algorithms

    International Nuclear Information System (INIS)

    Dinev, D.

    1996-01-01

    Several new algorithms for sorting of dipole and/or quadrupole magnets in synchrotrons and storage rings are described. The algorithms make use of a combinatorial approach to the problem and belong to the class of random search algorithms. They use an appropriate metrization of the state space. The phase-space distortion (smear) is used as a goal function. Computational experiments for the case of the JINR-Dubna superconducting heavy ion synchrotron NUCLOTRON have shown a significant reduction of the phase-space distortion after the magnet sorting. (orig.)

  11. Disposal of leached residual in heap leaching by neutralization

    International Nuclear Information System (INIS)

    Wang Jingmin

    1993-01-01

    The disposal results of leached residual with lime are described. Using the ratio of residual to lime being 100 : 1 the ideal disposal results were obtained with the effluent of the neutralized residual close to neutral

  12. Radiation effects on residual voltage of polyethylene films

    International Nuclear Information System (INIS)

    Kyokane, Jun; Park, Dae-Hee; Yoshino, Katsumi.

    1986-01-01

    It has recently been pointed out that diagnosis of deterioration in insulating materials for electric cables used in nuclear power plants and outer space (communications satellite in particular) can be effectively performed based on measurements of residual voltage. In the present study, polyethylene films are irradiated with γ-rays or electron beam to examine the changes in residual voltage characteristics. Irradiation of electron beam and γ-rays are carried out to a dose of 0 - 90 Mrad and 0 - 100 Mrad, respectively. Measurements are made of the dependence of residual voltage on applied voltage, electron beam and γ-ray irradiation, annealing temperature and annealing time. Results show that carriers, which are once trapped after being released from the electrode, move within the material after the opening of the circuit to produce resiual voltage. The residual voltage increases with increasing dose of electron beam or γ-ray and levels off at high dose. Residual voltage is increased about several times by either electron beam or γ-rays, but electron beam tends to cause greater residual voltage than γ-ray. Polyethylene films irradiated with electron beam can recover upon annealing. It is concluded from observations made that residual voltage has close relations with defects in molecular structures caused by radiations, particularly the breaking of backbone chains and alteration in superstructures. (Nogami, K.)

  13. An iterative bidirectional heuristic placement algorithm for solving the two-dimensional knapsack packing problem

    Science.gov (United States)

    Shiangjen, Kanokwatt; Chaijaruwanich, Jeerayut; Srisujjalertwaja, Wijak; Unachak, Prakarn; Somhom, Samerkae

    2018-02-01

    This article presents an efficient heuristic placement algorithm, namely, a bidirectional heuristic placement, for solving the two-dimensional rectangular knapsack packing problem. The heuristic demonstrates ways to maximize space utilization by fitting the appropriate rectangle from both sides of the wall of the current residual space layer by layer. The iterative local search along with a shift strategy is developed and applied to the heuristic to balance the exploitation and exploration tasks in the solution space without the tuning of any parameters. The experimental results on many scales of packing problems show that this approach can produce high-quality solutions for most of the benchmark datasets, especially for large-scale problems, within a reasonable duration of computational time.

  14. Composite Differential Search Algorithm

    Directory of Open Access Journals (Sweden)

    Bo Liu

    2014-01-01

    Full Text Available Differential search algorithm (DS is a relatively new evolutionary algorithm inspired by the Brownian-like random-walk movement which is used by an organism to migrate. It has been verified to be more effective than ABC, JDE, JADE, SADE, EPSDE, GSA, PSO2011, and CMA-ES. In this paper, we propose four improved solution search algorithms, namely “DS/rand/1,” “DS/rand/2,” “DS/current to rand/1,” and “DS/current to rand/2” to search the new space and enhance the convergence rate for the global optimization problem. In order to verify the performance of different solution search methods, 23 benchmark functions are employed. Experimental results indicate that the proposed algorithm performs better than, or at least comparable to, the original algorithm when considering the quality of the solution obtained. However, these schemes cannot still achieve the best solution for all functions. In order to further enhance the convergence rate and the diversity of the algorithm, a composite differential search algorithm (CDS is proposed in this paper. This new algorithm combines three new proposed search schemes including “DS/rand/1,” “DS/rand/2,” and “DS/current to rand/1” with three control parameters using a random method to generate the offspring. Experiment results show that CDS has a faster convergence rate and better search ability based on the 23 benchmark functions.

  15. Artifact removal algorithms for stroke detection using a multistatic MIST beamforming algorithm.

    Science.gov (United States)

    Ricci, E; Di Domenico, S; Cianca, E; Rossi, T

    2015-01-01

    Microwave imaging (MWI) has been recently proved as a promising imaging modality for low-complexity, low-cost and fast brain imaging tools, which could play a fundamental role to efficiently manage emergencies related to stroke and hemorrhages. This paper focuses on the UWB radar imaging approach and in particular on the processing algorithms of the backscattered signals. Assuming the use of the multistatic version of the MIST (Microwave Imaging Space-Time) beamforming algorithm, developed by Hagness et al. for the early detection of breast cancer, the paper proposes and compares two artifact removal algorithms. Artifacts removal is an essential step of any UWB radar imaging system and currently considered artifact removal algorithms have been shown not to be effective in the specific scenario of brain imaging. First of all, the paper proposes modifications of a known artifact removal algorithm. These modifications are shown to be effective to achieve good localization accuracy and lower false positives. However, the main contribution is the proposal of an artifact removal algorithm based on statistical methods, which allows to achieve even better performance but with much lower computational complexity.

  16. Law and Order in Algorithmics

    NARCIS (Netherlands)

    Fokkinga, M.M.

    1992-01-01

    An algorithm is the input-output effect of a computer program; mathematically, the notion of algorithm comes close to the notion of function. Just as arithmetic is the theory and practice of calculating with numbers, so is ALGORITHMICS the theory and practice of calculating with algorithms. Just as

  17. Coking of residue hydroprocessing catalysts

    Energy Technology Data Exchange (ETDEWEB)

    Gray, M.R.; Zhao, Y.X. [Alberta Univ., Edmonton, AB (Canada). Dept. of Chemical Engineering; McKnight, C.A. [Syncrude Canada Ltd., Edmonton, AB (Canada); Komar, D.A.; Carruthers, J.D. [Cytec Industries Inc., Stamford, CT (United States)

    1997-11-01

    One of the major causes of deactivation of Ni/Mo and Co/Mo sulfide catalysts for hydroprocessing of heavy petroleum and bitumen fractions is coke deposition. The composition and amount of coke deposited on residue hydroprocessing catalysts depends on the composition of the liquid phase of the reactor. In the Athabasca bitumen, the high molecular weight components encourage coke deposition at temperatures of 430 to 440 degrees C and at pressures of 10 to 20 MPa hydrogen pressure. A study was conducted to determine which components in the heavy residual oil fraction were responsible for coking of catalysts. Seven samples of Athabasca vacuum residue were prepared by supercritical fluid extraction with pentane before being placed in the reactor. Carbon content and hydrodesulfurization activity was measured. It was concluded that the deposition of coke depended on the presence of asphaltenes and not on other compositional variables such as content of nitrogen, aromatic carbon or vanadium.

  18. Algorithms in Algebraic Geometry

    CERN Document Server

    Dickenstein, Alicia; Sommese, Andrew J

    2008-01-01

    In the last decade, there has been a burgeoning of activity in the design and implementation of algorithms for algebraic geometric computation. Some of these algorithms were originally designed for abstract algebraic geometry, but now are of interest for use in applications and some of these algorithms were originally designed for applications, but now are of interest for use in abstract algebraic geometry. The workshop on Algorithms in Algebraic Geometry that was held in the framework of the IMA Annual Program Year in Applications of Algebraic Geometry by the Institute for Mathematics and Its

  19. Explaining algorithms using metaphors

    CERN Document Server

    Forišek, Michal

    2013-01-01

    There is a significant difference between designing a new algorithm, proving its correctness, and teaching it to an audience. When teaching algorithms, the teacher's main goal should be to convey the underlying ideas and to help the students form correct mental models related to the algorithm. This process can often be facilitated by using suitable metaphors. This work provides a set of novel metaphors identified and developed as suitable tools for teaching many of the 'classic textbook' algorithms taught in undergraduate courses worldwide. Each chapter provides exercises and didactic notes fo

  20. Portfolios of quantum algorithms.

    Science.gov (United States)

    Maurer, S M; Hogg, T; Huberman, B A

    2001-12-17

    Quantum computation holds promise for the solution of many intractable problems. However, since many quantum algorithms are stochastic in nature they can find the solution of hard problems only probabilistically. Thus the efficiency of the algorithms has to be characterized by both the expected time to completion and the associated variance. In order to minimize both the running time and its uncertainty, we show that portfolios of quantum algorithms analogous to those of finance can outperform single algorithms when applied to the NP-complete problems such as 3-satisfiability.

  1. Bloqueio neuromuscular residual após o uso de rocurônio ou cisatracúrio Bloqueo neuromuscular residual después del uso de rocuronio o cisatracúrio Residual neuromuscular block after rocuronium or cisatracurium

    Directory of Open Access Journals (Sweden)

    Bruno Salomé de Morais

    2005-12-01

    Full Text Available JUSTIFICATIVA E OBJETIVOS: O bloqueio neuromuscular residual na sala de recuperação pós-anestésica (SRPA é um fenômeno que pode aumentar a morbidade pós-operatória, com incidência variando entre 0% e 93%. O objetivo deste estudo foi avaliar a incidência do bloqueio neuromuscular residual na SRPA. MÉTODO: Foram estudados 93 pacientes submetidos à cirurgia geral com o uso de cisatracúrio ou rocurônio. Após a admissão na SRPA foi realizada a monitorização objetiva da função neuromuscular (aceleromiografia - TOF GUARD. O bloqueio neuromuscular residual foi definido como SQE JUSTIFICATIVA Y OBJETIVOS: El bloqueo neuromuscular residual en la sala de recuperación posanestésica (SRPA es un fenómeno que puede aumentar la morbidez posoperatoria, con incidencia variando entre 0% y 93%. La finalidad de este estudio fue evaluar la incidencia del bloqueo neuromuscular residual en la SRPA. MÉTODO: Fueron estudiados 93 pacientes sometidos a cirugía general con el uso de cisatracúrio o rocuronio. Después de la admisión en la SRPA fue realizada la monitorización objetiva de la función neuromuscular (aceleromiografia - TOF-GUARD. El bloqueo neuromuscular residual fue definido como TOF BACKGROUND AND OBJECTIVES: Residual neuromuscular block in the post-anesthetic recovery unit (PACU may increase postoperative morbidity from 0% to 93%. This study aimed at evaluating the incidence of residual neuromuscular block in the PACU. METHODS: Participated in this study 93 patients submitted to general anesthesia with cisatracurium or rocuronium. After PACU admission, neuromuscular function was objectively monitored (acceleromyography - TOF GUARD. Residual neuromuscular block was defined as TOF < 0.9. RESULTS: From 93 patients, 53 received cisatracurium and 40 rocuronium. Demographics, procedure length and the use of antagonists were comparable between groups. Residual neuromuscular block was 32% in subgroup C (cisatracurium and 30% in subgroup R

  2. Optimization of heat-liberating batches for ash residue stabilization

    International Nuclear Information System (INIS)

    Karlina, O.K.; Varlackova, G.A.; Ojovan, M.I.; Tivansky, V.M.; Dmitriev, S.A.

    1999-01-01

    The ash residue obtained after incineration of solid radioactive waste is a dusting poly-dispersed powder like material that contains radioactive nuclides ( 137 Cs, 90 Sr, 239 Pu, hor ( ellipsis)). Specific radioactivity of the ash can be about 10 5 --10 7 Bq/kg. In order to dispose of the ash, residue shall be stabilized by producing a monolith material. The ash residue can be either vitrified or stabilized into a ceramic matrix. For this purpose the ash residue is mixed with fluxing agents followed by melting of obtained composition in the different type melters. As a rule this requires both significant energy consumption and complex melting equipment. A stabilization technology of ash residue was proposed recently by using heat liberating batches-compositions with redox properties. The ash residue is melted due to exothermic chemical reactions in the mixture with heat-liberating batch that occur with considerable release of heat. Stabilization method has three stages: (1) preparation of a mixture of heating batch and ash residue with or without glass forming batch (frit); (2) ignition and combustion of mixed composition; (3) cooling (quenching) of obtained vitreous material. Combustion of mixed composition occurs in the form of propagation of reacting wave. The heat released during exothermic chemical reactions provides melting of ash residue components and production of glass-like phase. The final product consists of a glass like matrix with embedded crystalline inclusions of infusible ash residue components

  3. DNABIT Compress - Genome compression algorithm.

    Science.gov (United States)

    Rajarajeswari, Pothuraju; Apparao, Allam

    2011-01-22

    Data compression is concerned with how information is organized in data. Efficient storage means removal of redundancy from the data being stored in the DNA molecule. Data compression algorithms remove redundancy and are used to understand biologically important molecules. We present a compression algorithm, "DNABIT Compress" for DNA sequences based on a novel algorithm of assigning binary bits for smaller segments of DNA bases to compress both repetitive and non repetitive DNA sequence. Our proposed algorithm achieves the best compression ratio for DNA sequences for larger genome. Significantly better compression results show that "DNABIT Compress" algorithm is the best among the remaining compression algorithms. While achieving the best compression ratios for DNA sequences (Genomes),our new DNABIT Compress algorithm significantly improves the running time of all previous DNA compression programs. Assigning binary bits (Unique BIT CODE) for (Exact Repeats, Reverse Repeats) fragments of DNA sequence is also a unique concept introduced in this algorithm for the first time in DNA compression. This proposed new algorithm could achieve the best compression ratio as much as 1.58 bits/bases where the existing best methods could not achieve a ratio less than 1.72 bits/bases.

  4. Residual gravimetric method to measure nebulizer output.

    Science.gov (United States)

    Vecellio None, Laurent; Grimbert, Daniel; Bordenave, Joelle; Benoit, Guy; Furet, Yves; Fauroux, Brigitte; Boissinot, Eric; De Monte, Michele; Lemarié, Etienne; Diot, Patrice

    2004-01-01

    The aim of this study was to assess a residual gravimetric method based on weighing dry filters to measure the aerosol output of nebulizers. This residual gravimetric method was compared to assay methods based on spectrophotometric measurement of terbutaline (Bricanyl, Astra Zeneca, France), high-performance liquid chromatography (HPLC) measurement of tobramycin (Tobi, Chiron, U.S.A.), and electrochemical measurements of NaF (as defined by the European standard). Two breath-enhanced jet nebulizers, one standard jet nebulizer, and one ultrasonic nebulizer were tested. Output produced by the residual gravimetric method was calculated by weighing the filters both before and after aerosol collection and by filter drying corrected by the proportion of drug contained in total solute mass. Output produced by the electrochemical, spectrophotometric, and HPLC methods was determined after assaying the drug extraction filter. The results demonstrated a strong correlation between the residual gravimetric method (x axis) and assay methods (y axis) in terms of drug mass output (y = 1.00 x -0.02, r(2) = 0.99, n = 27). We conclude that a residual gravimetric method based on dry filters, when validated for a particular agent, is an accurate way of measuring aerosol output.

  5. Incorporation feasibility of leather residues in bricks

    Energy Technology Data Exchange (ETDEWEB)

    Aguiar, J.B. [Minho Univ. (Portugal). Civil Engineering Dept.; Valente, A.; Pires, M.J. [Inst. of Development and Innovation Technology of Minho, Braga (Portugal); Tavares, T. [Biological Engineering Dept., Univ. of Minho, Braga (Portugal)

    2002-07-01

    The footwear industry has strips of leather as one of its by-products. These leather residues, due to their high chromium content, can be regarded as a threat to the environment, particularly if no care is taken with their disposal. With the incorporation of the residues in ceramic products, after trituration, is possible to neutralise the eventual toxicity of chromium. In a laboratory study we produced prismatic bricks using clay from the region and incorporating 1, 3 and 5% (by mass) of leather residues. This corresponds at about 20, 60 and 100% (by apparent volume). The moulds were filled up with paste and, in order to have some compactness, the ceramic paste was compressed with a spatula. After that, it began the process of drying and burning the bricks. They were tested to flexure, compression and leaching. The results showed that the toxicity of chromium disappeared in the bricks. The mechanical tests showed a decrease in strength for the specimens with leather residue. The compressive strength decreases about 22% for 1% of incorporation of leather residue. However, as bricks were lighter and more porous, we can expect that they are better for thermal isolation. (orig.)

  6. Assessment of pesticide residues on selected vegetables of Pakistan

    International Nuclear Information System (INIS)

    Khan, M.S.; Shah, M.M.

    2011-01-01

    The present study was conducted to determine the pesticide residues on selected summer vegetables. Five vegetables were grown with three replicates in a split plot randomized complete block design. Pesticides were sprayed on vegetables thrice at regular intervals each after 15 days. At maturity the pesticides residues were extracted from edible and leaf portions using anhydrous sodium sulfate and ethyl acetate while adsorption chromatography technique was used for cleanup. The extracts were subjected to high performance liquid chromatography (HPLC) for separation and analysis of the compounds. Significant differences (p<0.05) were found in the pesticides residues on edible portions whereas highly significant differences (p<0.001) were observed for the leafy portions. The residual level of cypermethrin was highest (16.2 mg kg/sup -1/) in edible portion of bitter gourd, while Lambdacyhalothrin and Mancozeb residues were detected high (4.50 mg kg/sup -1/, 6.26 mg kg/sup -1/) in edible portion of bitter gourd and Cucumber respectively. Cypermethrin residues were high (1.86 mg kg/sup -1/) in Okra leaves. Mancozeb and Lambdacyhalothrin residual level was high (1.23 mg kg/sup -1/, and 0.0002 mg kg/sup -1/) in chili and tomato leaves. Cypermethrin residues were readily detected in edible and leaf portion of the selected vegetables. (author)

  7. Stabilization of plutonium bearing residues at Lawrence Livermore National Laboratory

    International Nuclear Information System (INIS)

    Bronson, M.C.; Van Konynenburg, R.A.; Ebbinghaus, B.B.

    1995-01-01

    The US Department of Energy's (US DOE) Lawrence Livermore National Laboratory (LLNL) has plutonium holdings including metal, oxide and residue materials, all of which need stabilization of some type. Residue materials include calcined ash, calcined precipitates, pyrochemical salts, glove box sweepings, metallurgical samples, graphite, and pyrochemical ceramic crucibles. These residues are typical of residues stored throughout the US DOE plutonium sites. The stabilization process selected for each of these residues requires data on chemical impurities, physical attributes, and chemical forms of the plutonium. This paper outlines the characterization and stabilization of LLNL ash residues, pyrochemical salts, and graphite

  8. A verified LLL algorithm

    NARCIS (Netherlands)

    Divasón, Jose; Joosten, Sebastiaan; Thiemann, René; Yamada, Akihisa

    2018-01-01

    The Lenstra-Lenstra-Lovász basis reduction algorithm, also known as LLL algorithm, is an algorithm to find a basis with short, nearly orthogonal vectors of an integer lattice. Thereby, it can also be seen as an approximation to solve the shortest vector problem (SVP), which is an NP-hard problem,

  9. A Multi-Factor Analysis of Sustainable Agricultural Residue Removal Potential

    Energy Technology Data Exchange (ETDEWEB)

    Jared Abodeely; David Muth; Paul Adler; Eleanor Campbell; Kenneth Mark Bryden

    2012-10-01

    Agricultural residues have significant potential as a near term source of cellulosic biomass for bioenergy production, but sustainable removal of agricultural residues requires consideration of the critical roles that residues play in the agronomic system. Previous work has developed an integrated model to evaluate sustainable agricultural residue removal potential considering soil erosion, soil organic carbon, greenhouse gas emission, and long-term yield impacts of residue removal practices. The integrated model couples the environmental process models WEPS, RUSLE2, SCI, and DAYCENT. This study uses the integrated model to investigate the impact of interval removal practices in Boone County, Iowa, US. Residue removal of 4.5 Mg/ha was performed annually, bi-annually, and tri-annually and were compared to no residue removal. The study is performed at the soil type scale using a national soil survey database assuming a continuous corn rotation with reduced tillage. Results are aggregated across soil types to provide county level estimates of soil organic carbon changes and individual soil type soil organic matter content if interval residue removal were implemented. Results show interval residue removal is possible while improving soil organic matter. Implementation of interval removal practices provide greater increases in soil organic matter while still providing substantial residue for bioenergy production.

  10. SPECIAL LIBRARIES OF FRAGMENTS OF ALGORITHMIC NETWORKS TO AUTOMATE THE DEVELOPMENT OF ALGORITHMIC MODELS

    Directory of Open Access Journals (Sweden)

    V. E. Marley

    2015-01-01

    Full Text Available Summary. The concept of algorithmic models appeared from the algorithmic approach in which the simulated object, the phenomenon appears in the form of process, subject to strict rules of the algorithm, which placed the process of operation of the facility. Under the algorithmic model is the formalized description of the scenario subject specialist for the simulated process, the structure of which is comparable with the structure of the causal and temporal relationships between events of the process being modeled, together with all information necessary for its software implementation. To represent the structure of algorithmic models used algorithmic network. Normally, they were defined as loaded finite directed graph, the vertices which are mapped to operators and arcs are variables, bound by operators. The language of algorithmic networks has great features, the algorithms that it can display indifference the class of all random algorithms. In existing systems, automation modeling based on algorithmic nets, mainly used by operators working with real numbers. Although this reduces their ability, but enough for modeling a wide class of problems related to economy, environment, transport, technical processes. The task of modeling the execution of schedules and network diagrams is relevant and useful. There are many counting systems, network graphs, however, the monitoring process based analysis of gaps and terms of graphs, no analysis of prediction execution schedule or schedules. The library is designed to build similar predictive models. Specifying source data to obtain a set of projections from which to choose one and take it for a new plan.

  11. Automated detection of residual cells after sex-mismatched stem-cell transplantation – evidence for presence of disease-marker negative residual cells

    Directory of Open Access Journals (Sweden)

    Johannes Tilman

    2009-05-01

    Full Text Available Abstract Background A new chimerism analysis based on automated interphase fluorescence in situ hybridization (FISH evaluation was established to detect residual cells after allogene sex-mismatched bone marrow or blood stem-cell transplantation. Cells of 58 patients were characterized as disease-associated due to presence of a bcr/abl-gene-fusion or a trisomy 8 and/or a simultaneous hybridization of gonosome-specific centromeric probes. The automatic slide scanning platform Metafer with its module MetaCyte was used to analyse 3,000 cells per sample. Results Overall 454 assays of 58 patients were analyzed. 13 of 58 patients showed residual recipient cells at one stage of more than 4% and 12 of 58 showed residual recipient cells less than 4%, respectively. As to be expected, patients of the latter group were associated with a higher survival rate (48 vs. 34 month. In only two of seven patients with disease-marker positive residual cells between 0.1–1.3% a relapse was observed. Besides, disease-marker negative residual cells were found in two patients without relapse at a rate of 2.8% and 3.3%, respectively. Conclusion The definite origin and meaning of disease-marker negative residual cells is still unclear. Overall, with the presented automatic chimerism analysis of interphase FISH slides, a sensitive method for detection of disease-marker positive residual cells is on hand.

  12. Evaluation of Crops Sensitivity to Atrazine Soil Residual

    Directory of Open Access Journals (Sweden)

    E Izadi

    2012-02-01

    Full Text Available Abstract In order to study the sensitivity of 9 crops to atrazine soil residual, two separate experiments were conducted in field and greenhouse conditions. First experiment was conducted in a field with treated soil by atrazine based on split plot and the results evaluated in greenhouse conditions. Treatments in the field experiment included two organic manure application rates (0 and 50 t/ha as main plots and 2 atrazine application rates (2 and 4 kg/ha atrazine a.i. as sob plots. After corn harvesting soil was sampled at 0-15 cm surface layer in each plots in 15 points, after mixing the samples. Wheat, barley, sugar beet, pea, lens and colza planted in pots at greenhouse. Second experiment conducted in greenhouse conditions for evaluation of atrazine soil residual in completely randomized design. Treatments included atrazine soil residual concentrations (0, 0.2, 0.5, 1, 5, 10 and 15 mg/kg soil and crops included wheat, barley, sugar beet, pea, lens, rape, bean and tomato. Results showed that atrazine residue had no effect on crops growth in field experiment treated with atrazine. It seems that atrazine residue in filed soil is lower that its damage threshold for crops or maybe for its fast removal in field than in control conditions. But in bioassay experiment (greenhouse experiment crops response to atrazine residues were different. Results showed that onion and pea were most susceptible ant tolerant crops between studied species and based of ED50 parameter the other crops tolerance to total residue ranked as: pea< bean< lentil< sugar beet< tomato< barley< wheat< rape< onion. Keywords: Atrazine residue, Pea, Bean, Lentil, Sugar beet, Barley, Wheat, Rape, Tomato

  13. Group leaders optimization algorithm

    Science.gov (United States)

    Daskin, Anmer; Kais, Sabre

    2011-03-01

    We present a new global optimization algorithm in which the influence of the leaders in social groups is used as an inspiration for the evolutionary technique which is designed into a group architecture. To demonstrate the efficiency of the method, a standard suite of single and multi-dimensional optimization functions along with the energies and the geometric structures of Lennard-Jones clusters are given as well as the application of the algorithm on quantum circuit design problems. We show that as an improvement over previous methods, the algorithm scales as N 2.5 for the Lennard-Jones clusters of N-particles. In addition, an efficient circuit design is shown for a two-qubit Grover search algorithm which is a quantum algorithm providing quadratic speedup over the classical counterpart.

  14. Totally parallel multilevel algorithms

    Science.gov (United States)

    Frederickson, Paul O.

    1988-01-01

    Four totally parallel algorithms for the solution of a sparse linear system have common characteristics which become quite apparent when they are implemented on a highly parallel hypercube such as the CM2. These four algorithms are Parallel Superconvergent Multigrid (PSMG) of Frederickson and McBryan, Robust Multigrid (RMG) of Hackbusch, the FFT based Spectral Algorithm, and Parallel Cyclic Reduction. In fact, all four can be formulated as particular cases of the same totally parallel multilevel algorithm, which are referred to as TPMA. In certain cases the spectral radius of TPMA is zero, and it is recognized to be a direct algorithm. In many other cases the spectral radius, although not zero, is small enough that a single iteration per timestep keeps the local error within the required tolerance.

  15. The global Minmax k-means algorithm.

    Science.gov (United States)

    Wang, Xiaoyan; Bai, Yanping

    2016-01-01

    The global k -means algorithm is an incremental approach to clustering that dynamically adds one cluster center at a time through a deterministic global search procedure from suitable initial positions, and employs k -means to minimize the sum of the intra-cluster variances. However the global k -means algorithm sometimes results singleton clusters and the initial positions sometimes are bad, after a bad initialization, poor local optimal can be easily obtained by k -means algorithm. In this paper, we modified the global k -means algorithm to eliminate the singleton clusters at first, and then we apply MinMax k -means clustering error method to global k -means algorithm to overcome the effect of bad initialization, proposed the global Minmax k -means algorithm. The proposed clustering method is tested on some popular data sets and compared to the k -means algorithm, the global k -means algorithm and the MinMax k -means algorithm. The experiment results show our proposed algorithm outperforms other algorithms mentioned in the paper.

  16. Dosing algorithm to target a predefined AUC in patients with primary central nervous system lymphoma receiving high dose methotrexate.

    Science.gov (United States)

    Joerger, Markus; Ferreri, Andrés J M; Krähenbühl, Stephan; Schellens, Jan H M; Cerny, Thomas; Zucca, Emanuele; Huitema, Alwin D R

    2012-02-01

    There is no consensus regarding optimal dosing of high dose methotrexate (HDMTX) in patients with primary CNS lymphoma. Our aim was to develop a convenient dosing algorithm to target AUC(MTX) in the range between 1000 and 1100 µmol l(-1) h. A population covariate model from a pooled dataset of 131 patients receiving HDMTX was used to simulate concentration-time curves of 10,000 patients and test the efficacy of a dosing algorithm based on 24 h MTX plasma concentrations to target the prespecified AUC(MTX) . These data simulations included interindividual, interoccasion and residual unidentified variability. Patients received a total of four simulated cycles of HDMTX and adjusted MTX dosages were given for cycles two to four. The dosing algorithm proposes MTX dose adaptations ranging from +75% in patients with MTX C(24) 12 µmol l(-1). The proposed dosing algorithm resulted in a marked improvement of the proportion of patients within the AUC(MTX) target between 1000 and 1100 µmol l(-1) h (11% with standard MTX dose, 35% with the adjusted dose) and a marked reduction of the interindividual variability of MTX exposure. A simple and practical dosing algorithm for HDMTX has been developed based on MTX 24 h plasma concentrations, and its potential efficacy in improving the proportion of patients within a prespecified target AUC(MTX) and reducing the interindividual variability of MTX exposure has been shown by data simulations. The clinical benefit of this dosing algorithm should be assessed in patients with primary central nervous system lymphoma (PCNSL). © 2011 The Authors. British Journal of Clinical Pharmacology © 2011 The British Pharmacological Society.

  17. Agrochemical residues in rivers sediments, Poas, Costa Rica

    International Nuclear Information System (INIS)

    Masis, Federico; Valdez, Juan; Leon, Sandra; Coto, Tatiana

    2008-01-01

    The organophosphorus and organochlorine agrochemical residues distribution in sediments of 3 rivers located in an ornamental plant production area were analyzed in Poas canton, Alajuela, Costa Rica. The study comprised 8 months in order to assure 3 seasonal episodes: dry, transitional, and rainy seasons. Sediments were taken in 10 sampling stations along the rivers and characterized by a determination of their organic matter and texture. In 7 out of 10 sampling stations pesticide residues were detected in at least 1 of 4 samplings, but quantified only in 4 stations. Agrochemical residues evaluated included 21 organophosphorus and organochlorine pesticides; however, we found residues of only 3 organochlorine pesticides, due their high persistence in the sediment. Residues corresponded to PCNB (80-800 μg.kg -1 ), Endosulfan-β (40-50 μg.kg -1 ), and Endosulfan-α (90 μg.kg -1 ). Chlorothalonil was detected in only one sample. (author) [es

  18. Obtention of ceramic pigments with residue from electroplating

    International Nuclear Information System (INIS)

    Boss, A.; Kniess, C.T.; Aguiar, B.M. de; Prates, P.B.; Milanez, K.

    2011-01-01

    The incorporation of industrial residues in industrial processes opens up new business opportunities and reduces the volume of extraction of raw materials, preserving natural resources, which are limited. An important residue is the mud from galvanic industry, consisting of alkali and transition metals. According to NBR 10004/2004, this residue can be classified as Class I (hazardous), depending on the concentration of metals present in the mud. This paper proposes a method for reusing the residue from electroplating in ceramic pigments. The characterization of residual plating was obtained by chemical analysis, mineralogical analysis and pH measurements. The electroplating waste was incorporated in different percentages on a standard pigment formula of industrial ceramic, consisting mainly of Zn, Fe and Cr. The obtained pigments were applied in ceramic glazes to colorimetric and visual analysis, which showed good results with the addition of up to 15% of industrial waste. (author)

  19. [Residual pleural thickening in tuberculous pleuritis. Associated factors

    Science.gov (United States)

    Ruiz, E; Alegre, J; Alemán, C; Vizcaya, S; Armadans, L; Segura, R M; Andreu, J; Iglesias, D; Fernández de Sevilla, T

    2000-10-01

    To study the factors related to the development of residual pleural thickening in pleural tuberculosis. We studied 39 patients with tuberculous pleural effusion. A chest X-ray was taken of each patient at the end of treatment. The patients' medical histories, pleural fluid findings and diagnostic chest films were evaluated. Residual pleural thickening was defined as thickening that was visibly greater than 2 mm in the lower side portion of the chest film. Residual pleural thickening developed in 26% of patients and was found mainly in men (RR = 3.86). In no patients with Löwenstein-Jensen cultures positive for Mycobacterium tuberculosis did pleural complications develop. Residual pleural thickening is a common complication of tuberculous pleural effusion. Residual pleural thickening in tuberculous pleurisy occurs more often in men and older patients, and in cases in which pleural liquid culture is negative for M. tuberculosis.

  20. PENETAPAN KADAR RESIDU ORGANOKLORIN DAN TAKSIRAN RESIKO KESEHATAN MASYARAKAT TERHADAP RESIDU PESTISIDA ORGANOKLORIN PADA 10 KOMODITI PANGAN

    Directory of Open Access Journals (Sweden)

    Ani Isnawati

    2012-10-01

    Full Text Available Usaha untuk mendapatkan hasil pertanian yang meningkat tidak ekonomis jika tidak menggunakan pestisida untuk menanggulangi serangan hama tanaman. Selain memberikan keuntungan bagi petani, disisi lain residu pestisida dapat membahayakan konsumen dalam batas-batas tertentu. Pemakaian pestisida  organoklorin telah dilarang penggunaannya melalui Menteri Pertanian nomor 434.1/kpts/TP.270/7/2001 karena sifatnya yang persisten. Oleh karena itu untuk mengetahui sejauh mana residu pestistda organoklorin masih ada dikomoditi makanan, maka dilakukan penetapan kadar residu pestisida organoklorin dan melakukan perhitungan taksiran resiko terhadap kesehatan masyarakat. Sampel adalah 10 jenis makanan (beras, }agung, kacang panjang, pisang ambon, tahu tempe, daging sapi, daging ayam, ikan mas, ikan gabus yang paling banyak dikonsumsi masyarakat yang digolongkan sebagai makanan pokok, lauk, sayur dan buah. Tempat pengambilan sampel dipilih secara purposif dt wilayah propinsi Jawa Barat dari 3 kota besar, yaitu Bandung, Cirebon dan Serang. sepuluh jenis sampel diambil secara acak sederhana di beberapa kios di satu pasar tradisional dart tiap kota. Penetapan kadar residu organoklorin dilakukan dengan menggunakan kromatografi gas. Sampel diambil secara acak sederhana dibeberapa kios di satu pasar tradisional dari tiap kota. Hasil penelitian menunjukkan bahwa daging sapi yang berasal dari Bandung merupakan jenis makanan yang terdeteksi pestisida organoklorin, yaitu: alfa-endosulfon dan beta endosulfon dengan kadar masing-masing 0,0284 mg/kg dan 0,0249 mg/kg. Kadar yang didapat masih di bawah BMR (Batas Maksimum Residu dan nilai ADI (Acceptable Daily Intake untuk endosulfon, sehingga tidak berisiko terhadap masyarakat yang mengkonsumsinya.