Research on parallel algorithm for sequential pattern mining
Zhou, Lijuan; Qin, Bai; Wang, Yu; Hao, Zhongxiao
2008-03-01
Sequential pattern mining is the mining of frequent sequences related to time or other orders from the sequence database. Its initial motivation is to discover the laws of customer purchasing in a time section by finding the frequent sequences. In recent years, sequential pattern mining has become an important direction of data mining, and its application field has not been confined to the business database and has extended to new data sources such as Web and advanced science fields such as DNA analysis. The data of sequential pattern mining has characteristics as follows: mass data amount and distributed storage. Most existing sequential pattern mining algorithms haven't considered the above-mentioned characteristics synthetically. According to the traits mentioned above and combining the parallel theory, this paper puts forward a new distributed parallel algorithm SPP(Sequential Pattern Parallel). The algorithm abides by the principal of pattern reduction and utilizes the divide-and-conquer strategy for parallelization. The first parallel task is to construct frequent item sets applying frequent concept and search space partition theory and the second task is to structure frequent sequences using the depth-first search method at each processor. The algorithm only needs to access the database twice and doesn't generate the candidated sequences, which abates the access time and improves the mining efficiency. Based on the random data generation procedure and different information structure designed, this paper simulated the SPP algorithm in a concrete parallel environment and implemented the AprioriAll algorithm. The experiments demonstrate that compared with AprioriAll, the SPP algorithm had excellent speedup factor and efficiency.
Large-scale sequential quadratic programming algorithms
Energy Technology Data Exchange (ETDEWEB)
Eldersveld, S.K.
1992-09-01
The problem addressed is the general nonlinear programming problem: finding a local minimizer for a nonlinear function subject to a mixture of nonlinear equality and inequality constraints. The methods studied are in the class of sequential quadratic programming (SQP) algorithms, which have previously proved successful for problems of moderate size. Our goal is to devise an SQP algorithm that is applicable to large-scale optimization problems, using sparse data structures and storing less curvature information but maintaining the property of superlinear convergence. The main features are: 1. The use of a quasi-Newton approximation to the reduced Hessian of the Lagrangian function. Only an estimate of the reduced Hessian matrix is required by our algorithm. The impact of not having available the full Hessian approximation is studied and alternative estimates are constructed. 2. The use of a transformation matrix Q. This allows the QP gradient to be computed easily when only the reduced Hessian approximation is maintained. 3. The use of a reduced-gradient form of the basis for the null space of the working set. This choice of basis is more practical than an orthogonal null-space basis for large-scale problems. The continuity condition for this choice is proven. 4. The use of incomplete solutions of quadratic programming subproblems. Certain iterates generated by an active-set method for the QP subproblem are used in place of the QP minimizer to define the search direction for the nonlinear problem. An implementation of the new algorithm has been obtained by modifying the code MINOS. Results and comparisons with MINOS and NPSOL are given for the new algorithm on a set of 92 test problems.
A Trust-region-based Sequential Quadratic Programming Algorithm
DEFF Research Database (Denmark)
Henriksen, Lars Christian; Poulsen, Niels Kjølstad
This technical note documents the trust-region-based sequential quadratic programming algorithm used in other works by the authors. The algorithm seeks to minimize a convex nonlinear cost function subject to linear inequalty constraints and nonlinear equality constraints.......This technical note documents the trust-region-based sequential quadratic programming algorithm used in other works by the authors. The algorithm seeks to minimize a convex nonlinear cost function subject to linear inequalty constraints and nonlinear equality constraints....
Li, Wei; Yang, Zhen; Hu, Haifeng
2014-01-01
Graphical models have been widely applied in solving distributed inference problems in wireless networks. In this paper, we formulate the cooperative localization problem in a mobile network as an inference problem on a factor graph. Using a sequential schedule of message updates, a sequential uniformly reweighted sum-product algorithm (SURW-SPA) is developed for mobile localization problems. The proposed algorithm combines the distributed nature of belief propagation (BP) with the improved p...
Sequential unconstrained minimization algorithms for constrained optimization
International Nuclear Information System (INIS)
Byrne, Charles
2008-01-01
The problem of minimizing a function f(x):R J → R, subject to constraints on the vector variable x, occurs frequently in inverse problems. Even without constraints, finding a minimizer of f(x) may require iterative methods. We consider here a general class of iterative algorithms that find a solution to the constrained minimization problem as the limit of a sequence of vectors, each solving an unconstrained minimization problem. Our sequential unconstrained minimization algorithm (SUMMA) is an iterative procedure for constrained minimization. At the kth step we minimize the function G k (x)=f(x)+g k (x), to obtain x k . The auxiliary functions g k (x):D subset of R J → R + are nonnegative on the set D, each x k is assumed to lie within D, and the objective is to minimize the continuous function f:R J → R over x in the set C = D-bar, the closure of D. We assume that such minimizers exist, and denote one such by x-circumflex. We assume that the functions g k (x) satisfy the inequalities 0≤g k (x)≤G k-1 (x)-G k-1 (x k-1 ), for k = 2, 3, .... Using this assumption, we show that the sequence {(x k )} is decreasing and converges to f(x-circumflex). If the restriction of f(x) to D has bounded level sets, which happens if x-circumflex is unique and f(x) is closed, proper and convex, then the sequence {x k } is bounded, and f(x*)=f(x-circumflex), for any cluster point x*. Therefore, if x-circumflex is unique, x* = x-circumflex and {x k } → x-circumflex. When x-circumflex is not unique, convergence can still be obtained, in particular cases. The SUMMA includes, as particular cases, the well-known barrier- and penalty-function methods, the simultaneous multiplicative algebraic reconstruction technique (SMART), the proximal minimization algorithm of Censor and Zenios, the entropic proximal methods of Teboulle, as well as certain cases of gradient descent and the Newton–Raphson method. The proof techniques used for SUMMA can be extended to obtain related results
Sequential function approximation on arbitrarily distributed point sets
Wu, Kailiang; Xiu, Dongbin
2018-02-01
We present a randomized iterative method for approximating unknown function sequentially on arbitrary point set. The method is based on a recently developed sequential approximation (SA) method, which approximates a target function using one data point at each step and avoids matrix operations. The focus of this paper is on data sets with highly irregular distribution of the points. We present a nearest neighbor replacement (NNR) algorithm, which allows one to sample the irregular data sets in a near optimal manner. We provide mathematical justification and error estimates for the NNR algorithm. Extensive numerical examples are also presented to demonstrate that the NNR algorithm can deliver satisfactory convergence for the SA method on data sets with high irregularity in their point distributions.
Sequential bayes estimation algorithm with cubic splines on uniform meshes
International Nuclear Information System (INIS)
Hossfeld, F.; Mika, K.; Plesser-Walk, E.
1975-11-01
After outlining the principles of some recent developments in parameter estimation, a sequential numerical algorithm for generalized curve-fitting applications is presented combining results from statistical estimation concepts and spline analysis. Due to its recursive nature, the algorithm can be used most efficiently in online experimentation. Using computer-sumulated and experimental data, the efficiency and the flexibility of this sequential estimation procedure is extensively demonstrated. (orig.) [de
Efficient sequential and parallel algorithms for record linkage.
Mamun, Abdullah-Al; Mi, Tian; Aseltine, Robert; Rajasekaran, Sanguthevar
2014-01-01
Integrating data from multiple sources is a crucial and challenging problem. Even though there exist numerous algorithms for record linkage or deduplication, they suffer from either large time needs or restrictions on the number of datasets that they can integrate. In this paper we report efficient sequential and parallel algorithms for record linkage which handle any number of datasets and outperform previous algorithms. Our algorithms employ hierarchical clustering algorithms as the basis. A key idea that we use is radix sorting on certain attributes to eliminate identical records before any further processing. Another novel idea is to form a graph that links similar records and find the connected components. Our sequential and parallel algorithms have been tested on a real dataset of 1,083,878 records and synthetic datasets ranging in size from 50,000 to 9,000,000 records. Our sequential algorithm runs at least two times faster, for any dataset, than the previous best-known algorithm, the two-phase algorithm using faster computation of the edit distance (TPA (FCED)). The speedups obtained by our parallel algorithm are almost linear. For example, we get a speedup of 7.5 with 8 cores (residing in a single node), 14.1 with 16 cores (residing in two nodes), and 26.4 with 32 cores (residing in four nodes). We have compared the performance of our sequential algorithm with TPA (FCED) and found that our algorithm outperforms the previous one. The accuracy is the same as that of this previous best-known algorithm.
DEFF Research Database (Denmark)
Pontefisso, Alessandro; Zappalorto, Michele; Quaresimin, Marino
2016-01-01
In this work, a study of the Random Sequential Absorption (RSA) algorithm in the generation of nanoplatelet Volume Elements (VEs) is carried out. The effect of the algorithm input parameters on the reinforcement distribution is studied through the implementation of statistical tools, showing...... that the platelet distribution is systematically affected by these parameters. The consequence is that a parametric analysis of the VE input parameters may be biased by hidden differences in the filler distribution. The same statistical tools used in the analysis are implemented in a modified RSA algorithm...
Distributed Algorithms for Time Optimal Reachability Analysis
DEFF Research Database (Denmark)
Zhang, Zhengkui; Nielsen, Brian; Larsen, Kim Guldstrand
2016-01-01
. We propose distributed computing to accelerate time optimal reachability analysis. We develop five distributed state exploration algorithms, implement them in \\uppaal enabling it to exploit the compute resources of a dedicated model-checking cluster. We experimentally evaluate the implemented...... algorithms with four models in terms of their ability to compute near- or proven-optimal solutions, their scalability, time and memory consumption and communication overhead. Our results show that distributed algorithms work much faster than sequential algorithms and have good speedup in general.......Time optimal reachability analysis is a novel model based technique for solving scheduling and planning problems. After modeling them as reachability problems using timed automata, a real-time model checker can compute the fastest trace to the goal states which constitutes a time optimal schedule...
Systolic array processing of the sequential decoding algorithm
Chang, C. Y.; Yao, K.
1989-01-01
A systolic array processing technique is applied to implementing the stack algorithm form of the sequential decoding algorithm. It is shown that sorting, a key function in the stack algorithm, can be efficiently realized by a special type of systolic arrays known as systolic priority queues. Compared to the stack-bucket algorithm, this approach is shown to have the advantages that the decoding always moves along the optimal path, that it has a fast and constant decoding speed and that its simple and regular hardware architecture is suitable for VLSI implementation. Three types of systolic priority queues are discussed: random access scheme, shift register scheme and ripple register scheme. The property of the entries stored in the systolic priority queue is also investigated. The results are applicable to many other basic sorting type problems.
Articulated Human Motion Tracking Using Sequential Immune Genetic Algorithm
Directory of Open Access Journals (Sweden)
Yi Li
2013-01-01
Full Text Available We formulate human motion tracking as a high-dimensional constrained optimization problem. A novel generative method is proposed for human motion tracking in the framework of evolutionary computation. The main contribution is that we introduce immune genetic algorithm (IGA for pose optimization in latent space of human motion. Firstly, we perform human motion analysis in the learnt latent space of human motion. As the latent space is low dimensional and contents the prior knowledge of human motion, it makes pose analysis more efficient and accurate. Then, in the search strategy, we apply IGA for pose optimization. Compared with genetic algorithm and other evolutionary methods, its main advantage is the ability to use the prior knowledge of human motion. We design an IGA-based method to estimate human pose from static images for initialization of motion tracking. And we propose a sequential IGA (S-IGA algorithm for motion tracking by incorporating the temporal continuity information into the traditional IGA. Experimental results on different videos of different motion types show that our IGA-based pose estimation method can be used for initialization of motion tracking. The S-IGA-based motion tracking method can achieve accurate and stable tracking of 3D human motion.
International Nuclear Information System (INIS)
De Agostini, A.; Moretti, R.; Belletti, S.; Maira, G.; Magri, G.C.; Bestagno, M.
1992-01-01
The correction of organ movements in sequential radionuclide renography was done using an iterative algorithm that, by means of a set of rectangular regions of interest (ROIs), did not require any anatomical marker or manual elaboration of frames. The realignment programme here proposed is quite independent of the spatial and temporal distribution of activity and analyses the rotational movement in a simplified but reliable way. The position of the object inside a frame is evaluated by choosing the best ROI in a set of ROIs shifted 1 pixel around the central one. Statistical tests have to be fulfilled by the algorithm in order to activate the realignment procedure. Validation of the algorithm was done for different acquisition set-ups and organ movements. Results, summarized in Table 1, show that in about 90% of the stimulated experiments the algorithm is able to correct the movements of the object with a maximum error less of equal to 1 pixel limit. The usefulness of the realignment programme was demonstrated with sequential radionuclide renography as a typical clinical application. The algorithm-corrected curves of a 1-year-old patient were completely different from those obtained without a motion correction procedure. The algorithm may be applicable also to other types of scintigraphic examinations, besides functional imaging in which the realignment of frames of the dynamic sequence was an intrinsic demand. (orig.)
A Distributed Spanning Tree Algorithm
DEFF Research Database (Denmark)
Johansen, Karl Erik; Jørgensen, Ulla Lundin; Nielsen, Sven Hauge
We present a distributed algorithm for constructing a spanning tree for connected undirected graphs. Nodes correspond to processors and edges correspond to two-way channels. Each processor has initially a distinct identity and all processors perform the same algorithm. Computation as well...
A distributed spanning tree algorithm
DEFF Research Database (Denmark)
Johansen, Karl Erik; Jørgensen, Ulla Lundin; Nielsen, Svend Hauge
1988-01-01
We present a distributed algorithm for constructing a spanning tree for connected undirected graphs. Nodes correspond to processors and edges correspond to two way channels. Each processor has initially a distinct identity and all processors perform the same algorithm. Computation as well as comm...
Multimodal Estimation of Distribution Algorithms.
Yang, Qiang; Chen, Wei-Neng; Li, Yun; Chen, C L Philip; Xu, Xiang-Min; Zhang, Jun
2016-02-15
Taking the advantage of estimation of distribution algorithms (EDAs) in preserving high diversity, this paper proposes a multimodal EDA. Integrated with clustering strategies for crowding and speciation, two versions of this algorithm are developed, which operate at the niche level. Then these two algorithms are equipped with three distinctive techniques: 1) a dynamic cluster sizing strategy; 2) an alternative utilization of Gaussian and Cauchy distributions to generate offspring; and 3) an adaptive local search. The dynamic cluster sizing affords a potential balance between exploration and exploitation and reduces the sensitivity to the cluster size in the niching methods. Taking advantages of Gaussian and Cauchy distributions, we generate the offspring at the niche level through alternatively using these two distributions. Such utilization can also potentially offer a balance between exploration and exploitation. Further, solution accuracy is enhanced through a new local search scheme probabilistically conducted around seeds of niches with probabilities determined self-adaptively according to fitness values of these seeds. Extensive experiments conducted on 20 benchmark multimodal problems confirm that both algorithms can achieve competitive performance compared with several state-of-the-art multimodal algorithms, which is supported by nonparametric tests. Especially, the proposed algorithms are very promising for complex problems with many local optima.
Algorithm for Non-proportional Loading in Sequentially Linear Analysis
Yu, C.; Hoogenboom, P.C.J.; Rots, J.G.; Saouma, V.; Bolander, J.; Landis, E.
2016-01-01
Sequentially linear analysis (SLA) is an alternative to the Newton-Raphson method for analyzing the nonlinear behavior of reinforced concrete and masonry structures. In this paper SLA is extended to load cases that are applied one after the other, for example first dead load and then wind load. It
A fast exact sequential algorithm for the partial digest problem.
Abbas, Mostafa M; Bahig, Hazem M
2016-12-22
Restriction site analysis involves determining the locations of restriction sites after the process of digestion by reconstructing their positions based on the lengths of the cut DNA. Using different reaction times with a single enzyme to cut DNA is a technique known as a partial digestion. Determining the exact locations of restriction sites following a partial digestion is challenging due to the computational time required even with the best known practical algorithm. In this paper, we introduce an efficient algorithm to find the exact solution for the partial digest problem. The algorithm is able to find all possible solutions for the input and works by traversing the solution tree with a breadth-first search in two stages and deleting all repeated subproblems. Two types of simulated data, random and Zhang, are used to measure the efficiency of the algorithm. We also apply the algorithm to real data for the Luciferase gene and the E. coli K12 genome. Our algorithm is a fast tool to find the exact solution for the partial digest problem. The percentage of improvement is more than 75% over the best known practical algorithm for the worst case. For large numbers of inputs, our algorithm is able to solve the problem in a suitable time, while the best known practical algorithm is unable.
A fast and accurate online sequential learning algorithm for feedforward networks.
Liang, Nan-Ying; Huang, Guang-Bin; Saratchandran, P; Sundararajan, N
2006-11-01
In this paper, we develop an online sequential learning algorithm for single hidden layer feedforward networks (SLFNs) with additive or radial basis function (RBF) hidden nodes in a unified framework. The algorithm is referred to as online sequential extreme learning machine (OS-ELM) and can learn data one-by-one or chunk-by-chunk (a block of data) with fixed or varying chunk size. The activation functions for additive nodes in OS-ELM can be any bounded nonconstant piecewise continuous functions and the activation functions for RBF nodes can be any integrable piecewise continuous functions. In OS-ELM, the parameters of hidden nodes (the input weights and biases of additive nodes or the centers and impact factors of RBF nodes) are randomly selected and the output weights are analytically determined based on the sequentially arriving data. The algorithm uses the ideas of ELM of Huang et al. developed for batch learning which has been shown to be extremely fast with generalization performance better than other batch training methods. Apart from selecting the number of hidden nodes, no other control parameters have to be manually chosen. Detailed performance comparison of OS-ELM is done with other popular sequential learning algorithms on benchmark problems drawn from the regression, classification and time series prediction areas. The results show that the OS-ELM is faster than the other sequential algorithms and produces better generalization performance.
Efficient sequential and parallel algorithms for planted motif search.
Nicolae, Marius; Rajasekaran, Sanguthevar
2014-01-31
Motif searching is an important step in the detection of rare events occurring in a set of DNA or protein sequences. One formulation of the problem is known as (l,d)-motif search or Planted Motif Search (PMS). In PMS we are given two integers l and d and n biological sequences. We want to find all sequences of length l that appear in each of the input sequences with at most d mismatches. The PMS problem is NP-complete. PMS algorithms are typically evaluated on certain instances considered challenging. Despite ample research in the area, a considerable performance gap exists because many state of the art algorithms have large runtimes even for moderately challenging instances. This paper presents a fast exact parallel PMS algorithm called PMS8. PMS8 is the first algorithm to solve the challenging (l,d) instances (25,10) and (26,11). PMS8 is also efficient on instances with larger l and d such as (50,21). We include a comparison of PMS8 with several state of the art algorithms on multiple problem instances. This paper also presents necessary and sufficient conditions for 3 l-mers to have a common d-neighbor. The program is freely available at http://engr.uconn.edu/~man09004/PMS8/. We present PMS8, an efficient exact algorithm for Planted Motif Search. PMS8 introduces novel ideas for generating common neighborhoods. We have also implemented a parallel version for this algorithm. PMS8 can solve instances not solved by any previous algorithms.
Sequential Service Restoration for Unbalanced Distribution Systems and Microgrids
International Nuclear Information System (INIS)
Chen, Bo; Chen, Chen; Wang, Jianhui; Butler-Purry, Karen L.
2017-01-01
The resilience and reliability of modern power systems are threatened by increasingly severe weather events and cyber-physical security events. An effective restoration methodology is desired to optimally integrate emerging smart grid technologies and pave the way for developing self-healing smart grids. In this paper, a sequential service restoration (SSR) framework is proposed to generate restoration solutions for distribution systems and microgrids in the event of large-scale power outages. The restoration solution contains a sequence of control actions that properly coordinate switches, distributed generators, and switchable loads to form multiple isolated microgrids. The SSR can be applied for three-phase unbalanced distribution systems and microgrids and can adapt to various operation conditions. Mathematical models are introduced for three-phase unbalanced power flow, voltage regulators, transformers, and loads. Furthermore, the SSR problem is formulated as a mixed-integer linear programming model, and its effectiveness is evaluated via the modified IEEE 123 node test feeder.
Directory of Open Access Journals (Sweden)
Wodziński Marek
2017-06-01
Full Text Available This paper presents an alternative approach to the sequential data classification, based on traditional machine learning algorithms (neural networks, principal component analysis, multivariate Gaussian anomaly detector and finding the shortest path in a directed acyclic graph, using A* algorithm with a regression-based heuristic. Palm gestures were used as an example of the sequential data and a quadrocopter was the controlled object. The study includes creation of a conceptual model and practical construction of a system using the GPU to ensure the realtime operation. The results present the classification accuracy of chosen gestures and comparison of the computation time between the CPU- and GPU-based solutions.
Discrete Riccati equation solutions: Distributed algorithms
Directory of Open Access Journals (Sweden)
D. G. Lainiotis
1996-01-01
Full Text Available In this paper new distributed algorithms for the solution of the discrete Riccati equation are introduced. The algorithms are used to provide robust and computational efficient solutions to the discrete Riccati equation. The proposed distributed algorithms are theoretically interesting and computationally attractive.
Sequential and Parallel Algorithms for Finding a Maximum Convex Polygon
DEFF Research Database (Denmark)
Fischer, Paul
1997-01-01
This paper investigates the problem where one is given a finite set of n points in the plane each of which is labeled either ?positive? or ?negative?. We consider bounded convex polygons, the vertices of which are positive points and which do not contain any negative point. It is shown how...... such a polygon which is maximal with respect to area can be found in time O(n³ log n). With the same running time one can also find such a polygon which contains a maximum number of positive points. If, in addition, the number of vertices of the polygon is restricted to be at most M, then the running time...... becomes O(M n³ log n). It is also shown how to find a maximum convex polygon which contains a given point in time O(n³ log n). Two parallel algorithms for the basic problem are also presented. The first one runs in time O(n log n) using O(n²) processors, the second one has polylogarithmic time but needs O...
Efficient sequential and parallel algorithms for finding edit distance based motifs.
Pal, Soumitra; Xiao, Peng; Rajasekaran, Sanguthevar
2016-08-18
Motif search is an important step in extracting meaningful patterns from biological data. The general problem of motif search is intractable and there is a pressing need to develop efficient, exact and approximation algorithms to solve this problem. In this paper, we present several novel, exact, sequential and parallel algorithms for solving the (l,d) Edit-distance-based Motif Search (EMS) problem: given two integers l,d and n biological strings, find all strings of length l that appear in each input string with atmost d errors of types substitution, insertion and deletion. One popular technique to solve the problem is to explore for each input string the set of all possible l-mers that belong to the d-neighborhood of any substring of the input string and output those which are common for all input strings. We introduce a novel and provably efficient neighborhood exploration technique. We show that it is enough to consider the candidates in neighborhood which are at a distance exactly d. We compactly represent these candidate motifs using wildcard characters and efficiently explore them with very few repetitions. Our sequential algorithm uses a trie based data structure to efficiently store and sort the candidate motifs. Our parallel algorithm in a multi-core shared memory setting uses arrays for storing and a novel modification of radix-sort for sorting the candidate motifs. The algorithms for EMS are customarily evaluated on several challenging instances such as (8,1), (12,2), (16,3), (20,4), and so on. The best previously known algorithm, EMS1, is sequential and in estimated 3 days solves up to instance (16,3). Our sequential algorithms are more than 20 times faster on (16,3). On other hard instances such as (9,2), (11,3), (13,4), our algorithms are much faster. Our parallel algorithm has more than 600 % scaling performance while using 16 threads. Our algorithms have pushed up the state-of-the-art of EMS solvers and we believe that the techniques introduced in
Weighted-Bit-Flipping-Based Sequential Scheduling Decoding Algorithms for LDPC Codes
Directory of Open Access Journals (Sweden)
Qing Zhu
2013-01-01
Full Text Available Low-density parity-check (LDPC codes can be applied in a lot of different scenarios such as video broadcasting and satellite communications. LDPC codes are commonly decoded by an iterative algorithm called belief propagation (BP over the corresponding Tanner graph. The original BP updates all the variable-nodes simultaneously, followed by all the check-nodes simultaneously as well. We propose a sequential scheduling algorithm based on weighted bit-flipping (WBF algorithm for the sake of improving the convergence speed. Notoriously, WBF is a low-complexity and simple algorithm. We combine it with BP to obtain advantages of these two algorithms. Flipping function used in WBF is borrowed to determine the priority of scheduling. Simulation results show that it can provide a good tradeoff between FER performance and computation complexity for short-length LDPC codes.
Distribution Bottlenecks in Classification Algorithms
Zwartjes, G.J.; Havinga, Paul J.M.; Smit, Gerardus Johannes Maria; Hurink, Johann L.
2012-01-01
The abundance of data available on Wireless Sensor Networks makes online processing necessary. In industrial applications for example, the correct operation of equipment can be the point of interest while raw sampled data is of minor importance. Classication algorithms can be used to make state
Precise algorithm to generate random sequential adsorption of hard polygons at saturation
Zhang, G.
2018-04-01
Random sequential adsorption (RSA) is a time-dependent packing process, in which particles of certain shapes are randomly and sequentially placed into an empty space without overlap. In the infinite-time limit, the density approaches a "saturation" limit. Although this limit has attracted particular research interest, the majority of past studies could only probe this limit by extrapolation. We have previously found an algorithm to reach this limit using finite computational time for spherical particles and could thus determine the saturation density of spheres with high accuracy. In this paper, we generalize this algorithm to generate saturated RSA packings of two-dimensional polygons. We also calculate the saturation density for regular polygons of three to ten sides and obtain results that are consistent with previous, extrapolation-based studies.
Directory of Open Access Journals (Sweden)
Sheng Wanxing
2016-01-01
Full Text Available In allusion to the randomness of output power of distributed generation (DG, a reliability evaluation model based on sequential Monte Carlo simulation (SMCS for distribution system with DG is proposed. Operating states of the distribution system can be sampled by SMCS in chronological order thus the corresponding output power of DG can be generated. The proposed method has been tested on feeder F4 of IEEE-RBTS Bus 6. The results show that reliability evaluation of distribution system considering the uncertainty of output power of DG can be effectively implemented by SMCS.
Plane-Based Sampling for Ray Casting Algorithm in Sequential Medical Images
Lin, Lili; Chen, Shengyong; Shao, Yan; Gu, Zichun
2013-01-01
This paper proposes a plane-based sampling method to improve the traditional Ray Casting Algorithm (RCA) for the fast reconstruction of a three-dimensional biomedical model from sequential images. In the novel method, the optical properties of all sampling points depend on the intersection points when a ray travels through an equidistant parallel plan cluster of the volume dataset. The results show that the method improves the rendering speed at over three times compared with the conventional algorithm and the image quality is well guaranteed. PMID:23424608
A sequential quadratic programming algorithm using an incomplete solution of the subproblem
Energy Technology Data Exchange (ETDEWEB)
Murray, W. [Stanford Univ., CA (United States). Systems Optimization Lab.; Prieto, F.J. [Universidad `Carlos III` de Madrid (Spain). Dept. de Estadistica y Econometria
1993-05-01
We analyze sequential quadratic programming (SQP) methods to solve nonlinear constrained optimization problems that are more flexible in their definition than standard SQP methods. The type of flexibility introduced is motivated by the necessity to deviate from the standard approach when solving large problems. Specifically we no longer require a minimizer of the QP subproblem to be determined or particular Lagrange multiplier estimates to be used. Our main focus is on an SQP algorithm that uses a particular augmented Lagrangian merit function. New results are derived for this algorithm under weaker conditions than previously assumed; in particular, it is not assumed that the iterates lie on a compact set.
Fatigue reduction during aggregated and distributed sequential stimulation.
Bergquist, Austin J; Babbar, Vishvek; Ali, Saima; Popovic, Milos R; Masani, Kei
2017-08-01
Transcutaneous neuromuscular electrical stimulation (NMES) can generate muscle contractions for rehabilitation and exercise. However, NMES-evoked contractions are limited by fatigue when they are delivered "conventionally" (CONV) using a single active electrode. Researchers have developed "sequential" (SEQ) stimulation, involving rotation of pulses between multiple "aggregated" (AGGR-SEQ) or "distributed" (DISTR-SEQ) active electrodes, to reduce fatigue (torque-decline) by reducing motor unit discharge rates. The primary objective was to compare fatigue-related outcomes, "potentiation," "variability," and "efficiency" between CONV, AGGR-SEQ, and DISTR-SEQ stimulation of knee extensors in healthy participants. Torque and current were recorded during testing with fatiguing trains using each NMES type under isometric and isokinetic (180°/s) conditions. Compared with CONV stimulation, SEQ techniques reduced fatigue-related outcomes, increased potentiation, did not affect variability, and reduced efficiency. SEQ techniques hold promise for reducing fatigue during NMES-based rehabilitation and exercise; however, optimization is required to improve efficiency. Muscle Nerve 56: 271-281, 2017. © 2016 Wiley Periodicals, Inc.
Distribution agnostic structured sparsity recovery algorithms
Al-Naffouri, Tareq Y.; Masood, Mudassir
2013-01-01
We present an algorithm and its variants for sparse signal recovery from a small number of its measurements in a distribution agnostic manner. The proposed algorithm finds Bayesian estimate of a sparse signal to be recovered and at the same time
Privacy Preservation in Distributed Subgradient Optimization Algorithms
Lou, Youcheng; Yu, Lean; Wang, Shouyang
2015-01-01
Privacy preservation is becoming an increasingly important issue in data mining and machine learning. In this paper, we consider the privacy preserving features of distributed subgradient optimization algorithms. We first show that a well-known distributed subgradient synchronous optimization algorithm, in which all agents make their optimization updates simultaneously at all times, is not privacy preserving in the sense that the malicious agent can learn other agents' subgradients asymptotic...
Learning theory of distributed spectral algorithms
International Nuclear Information System (INIS)
Guo, Zheng-Chu; Lin, Shao-Bo; Zhou, Ding-Xuan
2017-01-01
Spectral algorithms have been widely used and studied in learning theory and inverse problems. This paper is concerned with distributed spectral algorithms, for handling big data, based on a divide-and-conquer approach. We present a learning theory for these distributed kernel-based learning algorithms in a regression framework including nice error bounds and optimal minimax learning rates achieved by means of a novel integral operator approach and a second order decomposition of inverse operators. Our quantitative estimates are given in terms of regularity of the regression function, effective dimension of the reproducing kernel Hilbert space, and qualification of the filter function of the spectral algorithm. They do not need any eigenfunction or noise conditions and are better than the existing results even for the classical family of spectral algorithms. (paper)
A distributed algorithm for machine learning
Chen, Shihong
2018-04-01
This paper considers a distributed learning problem in which a group of machines in a connected network, each learning its own local dataset, aim to reach a consensus at an optimal model, by exchanging information only with their neighbors but without transmitting data. A distributed algorithm is proposed to solve this problem under appropriate assumptions.
Synthesis of sequential control algorithms for pneumatic drives controlled by monostable valves
Directory of Open Access Journals (Sweden)
Ł. Dworzak
2009-07-01
Full Text Available Application of the Grafpol method [1] for synthesising sequential control algorithms for pneumatic drives controlled by monostable valves is presented. The developed principles simplify the MTS method of programming production processes in the scope of the memory realisation [2]. Thanks to this, time for synthesising the schematic equation can be significantly reduced in comparison to the network transformation method [3]. The designed schematic equation makes a ground for writing an application program of a PLC using any language defined in IEC 61131-3.
Distribution agnostic structured sparsity recovery algorithms
Al-Naffouri, Tareq Y.
2013-05-01
We present an algorithm and its variants for sparse signal recovery from a small number of its measurements in a distribution agnostic manner. The proposed algorithm finds Bayesian estimate of a sparse signal to be recovered and at the same time is indifferent to the actual distribution of its non-zero elements. Termed Support Agnostic Bayesian Matching Pursuit (SABMP), the algorithm also has the capability of refining the estimates of signal and required parameters in the absence of the exact parameter values. The inherent feature of the algorithm of being agnostic to the distribution of the data grants it the flexibility to adapt itself to several related problems. Specifically, we present two important extensions to this algorithm. One extension handles the problem of recovering sparse signals having block structures while the other handles multiple measurement vectors to jointly estimate the related unknown signals. We conduct extensive experiments to show that SABMP and its variants have superior performance to most of the state-of-the-art algorithms and that too at low-computational expense. © 2013 IEEE.
International Nuclear Information System (INIS)
Botet, R.
1996-01-01
A novel scaling of the multiplicity distributions is found in the shattering phase of the sequential fragmentation process with inhibition. The same scaling law is shown to hold in the percolation process. (author)
Fast, Distributed Algorithms in Deep Networks
2016-05-11
shallow networks, additional work will need to be done in order to allow for the application of ADMM to deep nets. The ADMM method allows for quick...Quock V Le, et al. Large scale distributed deep networks. In Advances in Neural Information Processing Systems, pages 1223–1231, 2012. [11] Ken-Ichi...A TRIDENT SCHOLAR PROJECT REPORT NO. 446 Fast, Distributed Algorithms in Deep Networks by Midshipman 1/C Ryan J. Burmeister, USN
A general algorithm for distributing information in a graph
Aji, Srinivas M.; McEliece, Robert J.
1997-01-01
We present a general “message-passing” algorithm for distributing information in a graph. This algorithm may help us to understand the approximate correctness of both the Gallager-Tanner-Wiberg algorithm, and the turbo-decoding algorithm.
Lee, Yu; Yu, Chanki; Lee, Sang Wook
2018-01-10
We present a sequential fitting-and-separating algorithm for surface reflectance components that separates individual dominant reflectance components and simultaneously estimates the corresponding bidirectional reflectance distribution function (BRDF) parameters from the separated reflectance values. We tackle the estimation of a Lafortune BRDF model, which combines a nonLambertian diffuse reflection and multiple specular reflectance components with a different specular lobe. Our proposed method infers the appropriate number of BRDF lobes and their parameters by separating and estimating each of the reflectance components using an interval analysis-based branch-and-bound method in conjunction with iterative K-ordered scale estimation. The focus of this paper is the estimation of the Lafortune BRDF model. Nevertheless, our proposed method can be applied to other analytical BRDF models such as the Cook-Torrance and Ward models. Experiments were carried out to validate the proposed method using isotropic materials from the Mitsubishi Electric Research Laboratories-Massachusetts Institute of Technology (MERL-MIT) BRDF database, and the results show that our method is superior to a conventional minimization algorithm.
Cooperated Bayesian algorithm for distributed scheduling problem
Institute of Scientific and Technical Information of China (English)
QIANG Lei; XIAO Tian-yuan
2006-01-01
This paper presents a new distributed Bayesian optimization algorithm (BOA) to overcome the efficiency problem when solving NP scheduling problems.The proposed approach integrates BOA into the co-evolutionary schema,which builds up a concurrent computing environment.A new search strategy is also introduced for local optimization process.It integrates the reinforcement learning(RL) mechanism into the BOA search processes,and then uses the mixed probability information from BOA (post-probability) and RL (pre-probability) to enhance the cooperation between different local controllers,which improves the optimization ability of the algorithm.The experiment shows that the new algorithm does better in both optimization (2.2%) and convergence (11.7%),compared with classic BOA.
International Nuclear Information System (INIS)
Woudstra, Evert; Heijmen, Ben J M; Storchi, Pascal R M
2008-01-01
Some time ago we developed and published a new deterministic algorithm (called Cycle) for automatic selection of beam orientations in radiotherapy. This algorithm is a plan generation process aiming at the prescribed PTV dose within hard dose and dose-volume constraints. The algorithm allows a large number of input orientations to be used and selects only the most efficient orientations, surviving the selection process. Efficiency is determined by a score function and is more or less equal to the extent of uninhibited access to the PTV for a specific beam during the selection process. In this paper we compare the capabilities of fast-simulated annealing (FSA) and Cycle for cases where local optima are supposed to be present. Five pancreas and five oesophagus cases previously treated in our institute were selected for this comparison. Plans were generated for FSA and Cycle, using the same hard dose and dose-volume constraints, and the largest possible achieved PTV doses as obtained from these algorithms were compared. The largest achieved PTV dose values were generally very similar for the two algorithms. In some cases FSA resulted in a slightly higher PTV dose than Cycle, at the cost of switching on substantially more beam orientations than Cycle. In other cases, when Cycle generated the solution with the highest PTV dose using only a limited number of non-zero weight beams, FSA seemed to have some difficulty in switching off the unfavourable directions. Cycle was faster than FSA, especially for large-dimensional feasible spaces. In conclusion, for the cases studied in this paper, we have found that despite the inherent drawback of sequential search as used by Cycle (where Cycle could probably get trapped in a local optimum), Cycle is nevertheless able to find comparable or sometimes slightly better treatment plans in comparison with FSA (which in theory finds the global optimum) especially in large-dimensional beam weight spaces
Yu, Liang; Antoni, Jerome; Leclere, Quentin; Jiang, Weikang
2017-11-01
Acoustical source reconstruction is a typical inverse problem, whose minimum frequency of reconstruction hinges on the size of the array and maximum frequency depends on the spacing distance between the microphones. For the sake of enlarging the frequency of reconstruction and reducing the cost of an acquisition system, Cyclic Projection (CP), a method of sequential measurements without reference, was recently investigated (JSV,2016,372:31-49). In this paper, the Propagation based Fast Iterative Shrinkage Thresholding Algorithm (Propagation-FISTA) is introduced, which improves CP in two aspects: (1) the number of acoustic sources is no longer needed and the only making assumption is that of a "weakly sparse" eigenvalue spectrum; (2) the construction of the spatial basis is much easier and adaptive to practical scenarios of acoustical measurements benefiting from the introduction of propagation based spatial basis. The proposed Propagation-FISTA is first investigated with different simulations and experimental setups and is next illustrated with an industrial case.
An algorithm for sequential tail value at risk for path-independent payoffs in a binomial tree
Roorda, Berend
2010-01-01
We present an algorithm that determines Sequential Tail Value at Risk (STVaR) for path-independent payoffs in a binomial tree. STVaR is a dynamic version of Tail-Value-at-Risk (TVaR) characterized by the property that risk levels at any moment must be in the range of risk levels later on. The
A Scalable Local Algorithm for Distributed Multivariate Regression
National Aeronautics and Space Administration — This paper offers a local distributed algorithm for multivariate regression in large peer-to-peer environments. The algorithm can be used for distributed...
An Efficient Local Algorithm for Distributed Multivariate Regression
National Aeronautics and Space Administration — This paper offers a local distributed algorithm for multivariate regression in large peer-to-peer environments. The algorithm is designed for distributed...
He, Chenlong; Feng, Zuren; Ren, Zhigang
2018-02-03
For Wireless Sensor Networks (WSNs), the Voronoi partition of a region is a challenging problem owing to the limited sensing ability of each sensor and the distributed organization of the network. In this paper, an algorithm is proposed for each sensor having a limited sensing range to compute its limited Voronoi cell autonomously, so that the limited Voronoi partition of the entire WSN is generated in a distributed manner. Inspired by Graham's Scan (GS) algorithm used to compute the convex hull of a point set, the limited Voronoi cell of each sensor is obtained by sequentially scanning two consecutive bisectors between the sensor and its neighbors. The proposed algorithm called the Boundary Scan (BS) algorithm has a lower computational complexity than the existing Range-Constrained Voronoi Cell (RCVC) algorithm and reaches the lower bound of the computational complexity of the algorithms used to solve the problem of this kind. Moreover, it also improves the time efficiency of a key step in the Adjust-Sensing-Radius (ASR) algorithm used to compute the exact Voronoi cell. Extensive numerical simulations are performed to demonstrate the correctness and effectiveness of the BS algorithm. The distributed realization of the BS combined with a localization algorithm in WSNs is used to justify the WSN nature of the proposed algorithm.
Energy Technology Data Exchange (ETDEWEB)
Serin, E.; Codel, G.; Mabhouti, H.; Cebe, M.; Sanli, E.; Pacaci, P.; Kucuk, N.; Kucukmorkoc, E.; Doyuran, M.; Canoglu, D.; Altinok, A.; Acar, H.; Caglar Ozkok, H. [Medipol University, Istanbul, Istanbul (Turkey)
2016-06-15
Purpose: In small field geometries, the electronic equilibrium can be lost, making it challenging for the dose-calculation algorithm to accurately predict the dose, especially in the presence of tissue heterogeneities. In this study, dosimetric accuracy of Monte Carlo (MC) advanced dose calculation and sequential algorithms of Multiplan treatment planning system were investigated for small radiation fields incident on homogeneous and heterogeneous geometries. Methods: Small open fields of fixed cones of Cyberknife M6 unit 100 to 500 mm2 were used for this study. The fields were incident on in house phantom containing lung, air, and bone inhomogeneities and also homogeneous phantom. Using the same film batch, the net OD to dose calibration curve was obtained using CK with the 60 mm fixed cone by delivering 0- 800 cGy. Films were scanned 48 hours after irradiation using an Epson 1000XL flatbed scanner. The dosimetric accuracy of MC and sequential algorithms in the presence of the inhomogeneities was compared against EBT3 film dosimetry Results: Open field tests in a homogeneous phantom showed good agreement between two algorithms and film measurement For MC algorithm, the minimum gamma analysis passing rates between measured and calculated dose distributions were 99.7% and 98.3% for homogeneous and inhomogeneous fields in the case of lung and bone respectively. For sequential algorithm, the minimum gamma analysis passing rates were 98.9% and 92.5% for for homogeneous and inhomogeneous fields respectively for used all cone sizes. In the case of the air heterogeneity, the differences were larger for both calculation algorithms. Overall, when compared to measurement, the MC had better agreement than sequential algorithm. Conclusion: The Monte Carlo calculation algorithm in the Multiplan treatment planning system is an improvement over the existing sequential algorithm. Dose discrepancies were observed for in the presence of air inhomogeneities.
International Nuclear Information System (INIS)
Serin, E.; Codel, G.; Mabhouti, H.; Cebe, M.; Sanli, E.; Pacaci, P.; Kucuk, N.; Kucukmorkoc, E.; Doyuran, M.; Canoglu, D.; Altinok, A.; Acar, H.; Caglar Ozkok, H.
2016-01-01
Purpose: In small field geometries, the electronic equilibrium can be lost, making it challenging for the dose-calculation algorithm to accurately predict the dose, especially in the presence of tissue heterogeneities. In this study, dosimetric accuracy of Monte Carlo (MC) advanced dose calculation and sequential algorithms of Multiplan treatment planning system were investigated for small radiation fields incident on homogeneous and heterogeneous geometries. Methods: Small open fields of fixed cones of Cyberknife M6 unit 100 to 500 mm2 were used for this study. The fields were incident on in house phantom containing lung, air, and bone inhomogeneities and also homogeneous phantom. Using the same film batch, the net OD to dose calibration curve was obtained using CK with the 60 mm fixed cone by delivering 0- 800 cGy. Films were scanned 48 hours after irradiation using an Epson 1000XL flatbed scanner. The dosimetric accuracy of MC and sequential algorithms in the presence of the inhomogeneities was compared against EBT3 film dosimetry Results: Open field tests in a homogeneous phantom showed good agreement between two algorithms and film measurement For MC algorithm, the minimum gamma analysis passing rates between measured and calculated dose distributions were 99.7% and 98.3% for homogeneous and inhomogeneous fields in the case of lung and bone respectively. For sequential algorithm, the minimum gamma analysis passing rates were 98.9% and 92.5% for for homogeneous and inhomogeneous fields respectively for used all cone sizes. In the case of the air heterogeneity, the differences were larger for both calculation algorithms. Overall, when compared to measurement, the MC had better agreement than sequential algorithm. Conclusion: The Monte Carlo calculation algorithm in the Multiplan treatment planning system is an improvement over the existing sequential algorithm. Dose discrepancies were observed for in the presence of air inhomogeneities.
DEFF Research Database (Denmark)
Rong, Aiying; Hakonen, Henri; Lahdelma, Risto
2009-01-01
efficiency of the plants. We introduce in this paper the DRDP-RSC algorithm, which is a dynamic regrouping based dynamic programming (DP) algorithm based on linear relaxation of the ON/OFF states of the units, sequential commitment of units in small groups. Relaxed states of the plants are used to reduce...... the dimension of the UC problem and dynamic regrouping is used to improve the solution quality. Numerical results based on real-life data sets show that this algorithm is efficient and optimal or near-optimal solutions with very small optimality gap are obtained....
Cardinality enhancement utilizing Sequential Algorithm (SeQ code in OCDMA system
Directory of Open Access Journals (Sweden)
Fazlina C. A. S.
2017-01-01
Full Text Available Optical Code Division Multiple Access (OCDMA has been important with increasing demand for high capacity and speed for communication in optical networks because of OCDMA technique high efficiency that can be achieved, hence fibre bandwidth is fully used. In this paper we will focus on Sequential Algorithm (SeQ code with AND detection technique using Optisystem design tool. The result revealed SeQ code capable to eliminate Multiple Access Interference (MAI and improve Bit Error Rate (BER, Phase Induced Intensity Noise (PIIN and orthogonally between users in the system. From the results, SeQ shows good performance of BER and capable to accommodate 190 numbers of simultaneous users contrast with existing code. Thus, SeQ code have enhanced the system about 36% and 111% of FCC and DCS code. In addition, SeQ have good BER performance 10-25 at 155 Mbps in comparison with 622 Mbps, 1 Gbps and 2 Gbps bit rate. From the plot graph, 155 Mbps bit rate is suitable enough speed for FTTH and LAN networks. Resolution can be made based on the superior performance of SeQ code. Thus, these codes will give an opportunity in OCDMA system for better quality of service in an optical access network for future generation's usage
Cardinality enhancement utilizing Sequential Algorithm (SeQ) code in OCDMA system
Fazlina, C. A. S.; Rashidi, C. B. M.; Rahman, A. K.; Aljunid, S. A.
2017-11-01
Optical Code Division Multiple Access (OCDMA) has been important with increasing demand for high capacity and speed for communication in optical networks because of OCDMA technique high efficiency that can be achieved, hence fibre bandwidth is fully used. In this paper we will focus on Sequential Algorithm (SeQ) code with AND detection technique using Optisystem design tool. The result revealed SeQ code capable to eliminate Multiple Access Interference (MAI) and improve Bit Error Rate (BER), Phase Induced Intensity Noise (PIIN) and orthogonally between users in the system. From the results, SeQ shows good performance of BER and capable to accommodate 190 numbers of simultaneous users contrast with existing code. Thus, SeQ code have enhanced the system about 36% and 111% of FCC and DCS code. In addition, SeQ have good BER performance 10-25 at 155 Mbps in comparison with 622 Mbps, 1 Gbps and 2 Gbps bit rate. From the plot graph, 155 Mbps bit rate is suitable enough speed for FTTH and LAN networks. Resolution can be made based on the superior performance of SeQ code. Thus, these codes will give an opportunity in OCDMA system for better quality of service in an optical access network for future generation's usage
An Improved Harmony Search Algorithm for Power Distribution Network Planning
Directory of Open Access Journals (Sweden)
Wei Sun
2015-01-01
Full Text Available Distribution network planning because of involving many variables and constraints is a multiobjective, discrete, nonlinear, and large-scale optimization problem. Harmony search (HS algorithm is a metaheuristic algorithm inspired by the improvisation process of music players. HS algorithm has several impressive advantages, such as easy implementation, less adjustable parameters, and quick convergence. But HS algorithm still has some defects such as premature convergence and slow convergence speed. According to the defects of the standard algorithm and characteristics of distribution network planning, an improved harmony search (IHS algorithm is proposed in this paper. We set up a mathematical model of distribution network structure planning, whose optimal objective function is to get the minimum annual cost and constraint conditions are overload and radial network. IHS algorithm is applied to solve the complex optimization mathematical model. The empirical results strongly indicate that IHS algorithm can effectively provide better results for solving the distribution network planning problem compared to other optimization algorithms.
A formal analysis of a dynamic distributed spanning tree algorithm
Mooij, A.J.; Wesselink, J.W.
2003-01-01
Abstract. We analyze the spanning tree algorithm in the IEEE 1394.1 draft standard, which correctness has not previously been proved. This algorithm is a fully-dynamic distributed graph algorithm, which, in general, is hard to develop. The approach we use is to formally develop an algorithm that is
Seghouane, Abd-Krim; Iqbal, Asif
2017-09-01
Sequential dictionary learning algorithms have been successfully applied to functional magnetic resonance imaging (fMRI) data analysis. fMRI data sets are, however, structured data matrices with the notions of temporal smoothness in the column direction. This prior information, which can be converted into a constraint of smoothness on the learned dictionary atoms, has seldomly been included in classical dictionary learning algorithms when applied to fMRI data analysis. In this paper, we tackle this problem by proposing two new sequential dictionary learning algorithms dedicated to fMRI data analysis by accounting for this prior information. These algorithms differ from the existing ones in their dictionary update stage. The steps of this stage are derived as a variant of the power method for computing the SVD. The proposed algorithms generate regularized dictionary atoms via the solution of a left regularized rank-one matrix approximation problem where temporal smoothness is enforced via regularization through basis expansion and sparse basis expansion in the dictionary update stage. Applications on synthetic data experiments and real fMRI data sets illustrating the performance of the proposed algorithms are provided.
International Nuclear Information System (INIS)
Huppertz, A.; Charite Universitaetsklinikum Berlin; Schmidt, M.; Schoeffski, O.; Wagner, M.; Asbach, P.; Maurer, M.H.; Puettcher, O.; Strassburg, J.; Stoeckmann, F.
2010-01-01
Purpose: To compare the direct costs of two diagnostic algorithms for pretherapeutic TNM staging of rectal cancer. Materials and Methods: In a study including 33 patients (mean age: 62.5 years), the direct fixed and variable costs of a sequential multimodal algorithm (rectoscopy, endoscopic and abdominal ultrasound, chest X-ray, thoracic/abdominal CT in the case of positive findings in abdominal ultrasound or chest X-ray) were compared to those of a novel algorithm of rectoscopy followed by MRI using a whole-body scanner. MRI included T 2w sequences of the rectum, 3D T 1w sequences of the liver and chest after bolus injection of gadoxetic acid, and delayed phases of the liver. The personnel work times, material items, and work processes were tracked to the nearest minute by interviewing those responsible for the process (surgeon, gastroenterologist, two radiologists). The costs of labor and materials were determined from personnel reimbursement data and hospital accounting records. Fixed costs were determined from vendor pricing. Results: The mean MRI time was 55 min. CT was performed in 19 / 33 patients (57 %) causing an additional day of hospitalization (costs 374 Euro). The costs for equipment and material were higher for MRI compared to sequential algorithm (equipment 116 vs. 30 Euro; material 159 vs. 60 Euro per patient). The personnel costs were markedly lower for MRI (436 vs. 732 Euro per patient). Altogether, the absolute cost advantage of MRI was 31.3 % (711 vs. 1035 Euro for sequential algorithm). Conclusion: Substantial savings are achievable with the use of whole-body MRI for the preoperative TNM staging of patients with rectal cancer. (orig.)
Park, Sang Cheol; Leader, Joseph Ken; Tan, Jun; Lee, Guee Sang; Kim, Soo Hyung; Na, In Seop; Zheng, Bin
2011-01-01
Objective this article presents a new computerized scheme that aims to accurately and robustly separate left and right lungs on CT examinations. Methods we developed and tested a method to separate the left and right lungs using sequential CT information and a guided dynamic programming algorithm using adaptively and automatically selected start point and end point with especially severe and multiple connections. Results the scheme successfully identified and separated all 827 connections on the total 4034 CT images in an independent testing dataset of CT examinations. The proposed scheme separated multiple connections regardless of their locations, and the guided dynamic programming algorithm reduced the computation time to approximately 4.6% in comparison with the traditional dynamic programming and avoided the permeation of the separation boundary into normal lung tissue. Conclusions The proposed method is able to robustly and accurately disconnect all connections between left and right lungs and the guided dynamic programming algorithm is able to remove redundant processing. PMID:21412104
Strong-field non-sequential ionization: The vector momentum distribution of multiply charged Ne ions
International Nuclear Information System (INIS)
Rottke, H.; Trump, C.; Wittmann, M.; Korn, G.; Becker, W.; Hoffmann, K.; Sandner, W.; Moshammer, R.; Feuerstein, B.; Dorn, A.; Schroeter, C.D.; Ullrich, J.; Schmitt, W.
2000-01-01
COLTRIMS (COLd Target Recoil-Ion Momentum Spectroscopy) was used to measure the vector momentum distribution of Ne n+ (n=1,2,3) ions formed in ultrashort (30 fsec) high-intensity (≅10 15 W/cm 2 ) laser pulses with center wavelength at 795 nm. To a high degree of accuracy the length of the Ne n+ ion momentum vector is equal to the length of the total momentum vector of the n photoelectrons released, with both vectors pointing into opposite directions. At a light intensity where non-sequential ionization of the atom dominates the Ne 2+ and Ne 3+ momentum distributions show distinct maxima at 4.0 a.u. and 7.5 a.u. along the polarization axis of the linearly polarized light beam. First, this is a clear signature of non-sequential multiple ionization. Second, it indicates that instantaneous emission of two (or more) electrons at electric field strength maxima of the light wave can be ruled out as main mechanism of non-sequential strong-field multiple ionization. In contrast, this experimental result is in accordance with the kinematical constraints of the 'rescattering model'
An improved algorithm for connectivity analysis of distribution networks
International Nuclear Information System (INIS)
Kansal, M.L.; Devi, Sunita
2007-01-01
In the present paper, an efficient algorithm for connectivity analysis of moderately sized distribution networks has been suggested. Algorithm is based on generation of all possible minimal system cutsets. The algorithm is efficient as it identifies only the necessary and sufficient conditions of system failure conditions in n-out-of-n type of distribution networks. The proposed algorithm is demonstrated with the help of saturated and unsaturated distribution networks. The computational efficiency of the algorithm is justified by comparing the computational efforts with the previously suggested appended spanning tree (AST) algorithm. The proposed technique has the added advantage as it can be utilized for generation of system inequalities which is useful in reliability estimation of capacitated networks
Block Least Mean Squares Algorithm over Distributed Wireless Sensor Network
Directory of Open Access Journals (Sweden)
T. Panigrahi
2012-01-01
Full Text Available In a distributed parameter estimation problem, during each sampling instant, a typical sensor node communicates its estimate either by the diffusion algorithm or by the incremental algorithm. Both these conventional distributed algorithms involve significant communication overheads and, consequently, defeat the basic purpose of wireless sensor networks. In the present paper, we therefore propose two new distributed algorithms, namely, block diffusion least mean square (BDLMS and block incremental least mean square (BILMS by extending the concept of block adaptive filtering techniques to the distributed adaptation scenario. The performance analysis of the proposed BDLMS and BILMS algorithms has been carried out and found to have similar performances to those offered by conventional diffusion LMS and incremental LMS algorithms, respectively. The convergence analyses of the proposed algorithms obtained from the simulation study are also found to be in agreement with the theoretical analysis. The remarkable and interesting aspect of the proposed block-based algorithms is that their communication overheads per node and latencies are less than those of the conventional algorithms by a factor as high as the block size used in the algorithms.
Archimedean copula estimation of distribution algorithm based on artificial bee colony algorithm
Institute of Scientific and Technical Information of China (English)
Haidong Xu; Mingyan Jiang; Kun Xu
2015-01-01
The artificial bee colony (ABC) algorithm is a com-petitive stochastic population-based optimization algorithm. How-ever, the ABC algorithm does not use the social information and lacks the knowledge of the problem structure, which leads to in-sufficiency in both convergent speed and searching precision. Archimedean copula estimation of distribution algorithm (ACEDA) is a relatively simple, time-economic and multivariate correlated EDA. This paper proposes a novel hybrid algorithm based on the ABC algorithm and ACEDA cal ed Archimedean copula estima-tion of distribution based on the artificial bee colony (ACABC) algorithm. The hybrid algorithm utilizes ACEDA to estimate the distribution model and then uses the information to help artificial bees to search more efficiently in the search space. Six bench-mark functions are introduced to assess the performance of the ACABC algorithm on numerical function optimization. Experimen-tal results show that the ACABC algorithm converges much faster with greater precision compared with the ABC algorithm, ACEDA and the global best (gbest)-guided ABC (GABC) algorithm in most of the experiments.
A Sequential Convex Semidefinite Programming Algorithm for Multiple-Load Free Material Optimization
Czech Academy of Sciences Publication Activity Database
Stingl, M.; Kočvara, Michal; Leugering, G.
2009-01-01
Roč. 20, č. 1 (2009), s. 130-155 ISSN 1052-6234 R&D Projects: GA AV ČR IAA1075402 Grant - others:commision EU(XE) EU-FP6-30717 Institutional research plan: CEZ:AV0Z10750506 Keywords : structural optimization * material optimization * semidefinite programming * sequential convex programming Subject RIV: BA - General Mathematics Impact factor: 1.429, year: 2009
An improved single sensor parity space algorithm for sequential probability ratio test
Energy Technology Data Exchange (ETDEWEB)
Racz, A. [Hungarian Academy of Sciences, Budapest (Hungary). Atomic Energy Research Inst.
1995-12-01
In our paper we propose a modification of the single sensor parity algorithm in order to make the statistical properties of the generated residual determinable in advance. The algorithm is tested via computer simulated ramp failure at the temperature readings of the pressurizer. (author).
Economic Models and Algorithms for Distributed Systems
Neumann, Dirk; Altmann, Jorn; Rana, Omer F
2009-01-01
Distributed computing models for sharing resources such as Grids, Peer-to-Peer systems, or voluntary computing are becoming increasingly popular. This book intends to discover fresh avenues of research and amendments to existing technologies, aiming at the successful deployment of commercial distributed systems
A Fast Algorithm for Generating Permutation Distribution of Ranks in ...
African Journals Online (AJOL)
... function of the distribution of the ranks. This further gives insight into the permutation distribution of a rank statistics. The algorithm is implemented with the aid of the computer algebra system Mathematica. Key words: Combinatorics, generating function, permutation distribution, rank statistics, partitions, computer algebra.
Lee, Jin-Ho; Kim, Dong-Jin; Ahn, Byung-Koo
2015-06-01
The objectives of this study were to investigate the distribution of thallium in soils collected near suspected areas such as cement plants, active and closed mines, and smelters and to examine the extraction of thallium in the soils using 19 single chemical and sequential chemical extraction procedures. Thallium concentrations in soils near cement plants were distributed between 1.20 and 12.91 mg kg(-1). However, soils near mines and smelters contained relatively low thallium concentrations ranging from 0.18 to 1.09 mg kg(-1). Thallium extractability with 19 single chemical extractants from selected soils near cement plants ranged from 0.10% to 8.20% of the total thallium concentration. In particular, 1.0 M NH4Cl, 1.0 M (NH4)2SO4, and 1.0 M CH3COONH4 extracted more thallium than other extractants. Sequential fractionation results of thallium from different soils such as industrially and artificially contaminated soils varied with the soil properties, especially soil pH and the duration of thallium contamination.
Agent-based Algorithm for Spatial Distribution of Objects
Collier, Nathan
2012-06-02
In this paper we present an agent-based algorithm for the spatial distribution of objects. The algorithm is a generalization of the bubble mesh algorithm, initially created for the point insertion stage of the meshing process of the finite element method. The bubble mesh algorithm treats objects in space as bubbles, which repel and attract each other. The dynamics of each bubble are approximated by solving a series of ordinary differential equations. We present numerical results for a meshing application as well as a graph visualization application.
A distributed scheduling algorithm for heterogeneous real-time systems
Zeineldine, Osman; El-Toweissy, Mohamed; Mukkamala, Ravi
1991-01-01
Much of the previous work on load balancing and scheduling in distributed environments was concerned with homogeneous systems and homogeneous loads. Several of the results indicated that random policies are as effective as other more complex load allocation policies. The effects of heterogeneity on scheduling algorithms for hard real time systems is examined. A distributed scheduler specifically to handle heterogeneities in both nodes and node traffic is proposed. The performance of the algorithm is measured in terms of the percentage of jobs discarded. While a random task allocation is very sensitive to heterogeneities, the algorithm is shown to be robust to such non-uniformities in system components and load.
Real-world experimentation of distributed DSA network algorithms
DEFF Research Database (Denmark)
Tonelli, Oscar; Berardinelli, Gilberto; Tavares, Fernando Menezes Leitão
2013-01-01
such as a dynamic propagation environment, human presence impact and terminals mobility. This chapter focuses on the practical aspects related to the real world-experimentation with distributed DSA network algorithms over a testbed network. Challenges and solutions are extensively discussed, from the testbed design......The problem of spectrum scarcity in uncoordinated and/or heterogeneous wireless networks is the key aspect driving the research in the field of flexible management of frequency resources. In particular, distributed dynamic spectrum access (DSA) algorithms enable an efficient sharing...... to the setup of experiments. A practical example of experimentation process with a DSA algorithm is also provided....
Implementing Distributed Algorithms using Remote Procedure Call
Bal, H.E.; van Renesse, R.; Tanenbaum, A.S.
1987-01-01
Remote procedure call (RPC) is a simple yet powerful primitiv~ for communication and synchronization between distributed processes. A problem with RPC is that it tends to decrease the amount of parallelism in an application due to its synchronous nature. This paper shows how light-weight processes
Du, Tingsong; Hu, Yang; Ke, Xianting
2015-01-01
An improved quantum artificial fish swarm algorithm (IQAFSA) for solving distributed network programming considering distributed generation is proposed in this work. The IQAFSA based on quantum computing which has exponential acceleration for heuristic algorithm uses quantum bits to code artificial fish and quantum revolving gate, preying behavior, and following behavior and variation of quantum artificial fish to update the artificial fish for searching for optimal value. Then, we apply the proposed new algorithm, the quantum artificial fish swarm algorithm (QAFSA), the basic artificial fish swarm algorithm (BAFSA), and the global edition artificial fish swarm algorithm (GAFSA) to the simulation experiments for some typical test functions, respectively. The simulation results demonstrate that the proposed algorithm can escape from the local extremum effectively and has higher convergence speed and better accuracy. Finally, applying IQAFSA to distributed network problems and the simulation results for 33-bus radial distribution network system show that IQAFSA can get the minimum power loss after comparing with BAFSA, GAFSA, and QAFSA.
A digital elevation analysis: Spatially distributed flow apportioning algorithm
Energy Technology Data Exchange (ETDEWEB)
Kim, Sang-Hyun; Kim, Kyung-Hyun [Pusan National University, Pusan(Korea); Jung, Sun-Hee [Korea Environment Institute, (Korea)
2001-06-30
A flow determination algorithm is proposed for the distributed hydrologic model. The advantages of a single flow direction scheme and multiple flow direction schemes are selectively considered to address the drawbacks of existing algorithms. A spatially varied flow apportioning factor is introduced in order to accommodate the accumulated area from upslope cells. The channel initiation threshold area(CIT) concept is expanded and integrated into the spatially distributed flow apportioning algorithm in order to delineate a realistic channel network. An application of a field example suggests that the linearly distributed flow apportioning scheme provides some advantages over existing approaches, such as the relaxation of over-dissipation problems near channel cells, the connectivity feature of river cells, the continuity of saturated areas and the negligence of the optimization of few parameters in existing algorithms. The effects of grid sizes are explored spatially as well as statistically. (author). 28 refs., 7 figs.
International Nuclear Information System (INIS)
Tsurumaru, Toyohiro
2007-01-01
In this paper, we discuss the security of the differential-phase-shift quantum-key-distribution (DPSQKD) protocol by introducing an improved version of the so-called sequential attack, which was originally discussed by Waks et al. [Phys. Rev. A 73, 012344 (2006)]. Our attack differs from the original form of the sequential attack in that the attacker Eve modulates not only the phases but also the amplitude in the superposition of the single-photon states which she sends to the receiver. Concentrating especially on the 'discretized Gaussian' intensity modulation, we show that our attack is more effective than the individual attack, which had been the best attack up to present. As a result of this, the recent experiment with communication distance of 100 km reported by Diamanti et al. [Opt. Express 14, 13073 (2006)] turns out to be insecure. Moreover, it can be shown that in a practical experimental setup which is commonly used today, the communication distance achievable by the DPSQKD protocol is less than 95 km
Distributed Time Synchronization Algorithms and Opinion Dynamics
Manita, Anatoly; Manita, Larisa
2018-01-01
We propose new deterministic and stochastic models for synchronization of clocks in nodes of distributed networks. An external accurate time server is used to ensure convergence of the node clocks to the exact time. These systems have much in common with mathematical models of opinion formation in multiagent systems. There is a direct analogy between the time server/node clocks pair in asynchronous networks and the leader/follower pair in the context of social network models.
Lee, Tae Hoon; Hwang, Soon Oh; Choi, Hyun Jong; Jung, Yunho; Cha, Sang Woo; Chung, Il-Kwun; Moon, Jong Ho; Cho, Young Deok; Park, Sang-Heum; Kim, Sun-Joo
2014-02-17
Numerous clinical trials to improve the success rate of biliary access in difficult biliary cannulation (DBC) during ERCP have been reported. However, standard guidelines or sequential protocol analysis according to different methods are limited in place. We planned to investigate a sequential protocol to facilitate selective biliary access for DBC during ERCP. This prospective clinical study enrolled 711 patients with naïve papillae at a tertiary referral center. If wire-guided cannulation was deemed to have failed due to the DBC criteria, then according to the cannulation algorithm early precut fistulotomy (EPF; cannulation time > 5 min, papillary contacts > 5 times, or hook-nose-shaped papilla), double-guidewire cannulation (DGC; unintentional pancreatic duct cannulation ≥ 3 times), and precut after placement of a pancreatic stent (PPS; if DGC was difficult or failed) were performed sequentially. The main outcome measurements were the technical success, procedure outcomes, and complications. Initially, a total of 140 (19.7%) patients with DBC underwent EPF (n = 71) and DGC (n = 69). Then, in DGC group 36 patients switched to PPS due to difficulty criteria. The successful biliary cannulation rate was 97.1% (136/140; 94.4% [67/71] with EPF, 47.8% [33/69] with DGC, and 100% [36/36] with PPS; P EPF, 314.8 (65.2) seconds in DGC, and 706.0 (469.4) seconds in PPS (P EPF, DGC, and PPS may be safe and feasible for DBC. The use of EPF in selected DBC criteria, DGC in unintentional pancreatic duct cannulations, and PPS in failed or difficult DGC may facilitate successful biliary cannulation.
Directory of Open Access Journals (Sweden)
Thakral B
2006-01-01
Full Text Available Background: HIV/AIDS pandemic brought into focus the importance of safe blood donor pool. Aims: To analyze true seroprevalence of HIV infection in our blood donors and devise an algorithm for donor recall avoiding unnecessary referrals to voluntary counseling and testing centre (VCTC. Materials and Methods: 39,784 blood units were screened for anti-HIV 1/2 using ELISA immunoassay (IA-1. Samples which were repeat reactive on IA-1 were further tested using two different immunoassays (IA-2 and IA-3 and Western blot (WB. Based on results of these sequential IAs and WB, an algorithm for recall of true HIV seroreactive blood donors is suggested for countries like India where nucleic acid testing or p24 antigen assays are not mandatory and given the limited resources may not be feasible. Results: The anti-HIV seroreactivity by repeat IA-1, IA-2, IA-3 and WB were 0.16%, 0.11%, 0.098% and 0.07% respectively. Of the 44 IA-1 reactive samples, 95.2% (20/21 of the seroreactive samples by both IA-2 and IA-3 were also WB positive and 100% (6/6 of the non-reactive samples by these IAs were WB negative. IA signal/cutoff ratio was significantly low in biological false reactive donors. WB indeterminate results were largely due to non-specific reactivity to gag protein (p55. Conclusions: HIV seroreactivity by sequential immunoassays (IA-1, IA-2 and IA-3; comparable to WHO Strategy-III prior to donor recall results in decreased referral to VCTC as compared to single IA (WHO Strategy-I being followed currently in India. Moreover, this strategy will repose donor confidence in our blood transfusion services and strengthen voluntary blood donation program.
Design of multiple sequence alignment algorithms on parallel, distributed memory supercomputers.
Church, Philip C; Goscinski, Andrzej; Holt, Kathryn; Inouye, Michael; Ghoting, Amol; Makarychev, Konstantin; Reumann, Matthias
2011-01-01
The challenge of comparing two or more genomes that have undergone recombination and substantial amounts of segmental loss and gain has recently been addressed for small numbers of genomes. However, datasets of hundreds of genomes are now common and their sizes will only increase in the future. Multiple sequence alignment of hundreds of genomes remains an intractable problem due to quadratic increases in compute time and memory footprint. To date, most alignment algorithms are designed for commodity clusters without parallelism. Hence, we propose the design of a multiple sequence alignment algorithm on massively parallel, distributed memory supercomputers to enable research into comparative genomics on large data sets. Following the methodology of the sequential progressiveMauve algorithm, we design data structures including sequences and sorted k-mer lists on the IBM Blue Gene/P supercomputer (BG/P). Preliminary results show that we can reduce the memory footprint so that we can potentially align over 250 bacterial genomes on a single BG/P compute node. We verify our results on a dataset of E.coli, Shigella and S.pneumoniae genomes. Our implementation returns results matching those of the original algorithm but in 1/2 the time and with 1/4 the memory footprint for scaffold building. In this study, we have laid the basis for multiple sequence alignment of large-scale datasets on a massively parallel, distributed memory supercomputer, thus enabling comparison of hundreds instead of a few genome sequences within reasonable time.
Algorithm describing pressure distribution of non-contact TNT explosion
Directory of Open Access Journals (Sweden)
Radosław Kiciński
2014-12-01
Full Text Available [b]Abstract[/b]. The aim of this study is to develop a computational algorithm, describing the shock wave pressure distribution in the space induced by non-contact TNT explosion. The procedure describes pressure distribution on a damp surface of the hull. Simulations have been carried out using Abaqus/CAE. The study also shows the pressure waveform descriptions provided by various authors and presents them in charts. The formulated conclusions convince efficiency of the algorithm application.[b]Keywords:[/b] Underwater explosion, shock wave, CAE, TNT, Kobben class submarine
Combining soft decision algorithms and scale-sequential hypotheses pruning for object recognition
Energy Technology Data Exchange (ETDEWEB)
Kumar, V.P.; Manolakos, E.S. [Northeastern Univ., Boston, MA (United States)
1996-12-31
This paper describes a system that exploits the synergy of Hierarchical Mixture Density (HMD) estimation with multiresolution decomposition based hypothesis pruning to perform efficiently joint segmentation and labeling of partially occluded objects in images. First we present the overall structure of the HMD estimation algorithm in the form of a recurrent neural network which generates the posterior probabilities of the various hypotheses associated with the image. Then in order to reduce the large memory and computation requirement we propose a hypothesis pruning scheme making use of the orthonormal discrete wavelet transform for dimensionality reduction. We provide an intuitive justification for the validity of this scheme and present experimental results and performance analysis on real and synthetic images to verify our claims.
Cocco, Arturo; Serra, Giuseppe; Lentini, Andrea; Deliperi, Salvatore; Delrio, Gavino
2015-09-01
The within- and between-plant distribution of the tomato leafminer, Tuta absoluta (Meyrick), was investigated in order to define action thresholds based on leaf infestation and to propose enumerative and binomial sequential sampling plans for pest management applications in protected crops. The pest spatial distribution was aggregated between plants, and median leaves were the most suitable sample to evaluate the pest density. Action thresholds of 36 and 48%, 43 and 56% and 60 and 73% infested leaves, corresponding to economic thresholds of 1 and 3% damaged fruits, were defined for tomato cultivars with big, medium and small fruits respectively. Green's method was a more suitable enumerative sampling plan as it required a lower sampling effort. Binomial sampling plans needed lower average sample sizes than enumerative plans to make a treatment decision, with probabilities of error of sampling plan required 87 or 343 leaves to estimate the population density in extensive or intensive ecological studies respectively. Binomial plans would be more practical and efficient for control purposes, needing average sample sizes of 17, 20 and 14 leaves to take a pest management decision in order to avoid fruit damage higher than 1% in cultivars with big, medium and small fruits respectively. © 2014 Society of Chemical Industry.
Intelligent decision support algorithm for distribution system restoration.
Singh, Reetu; Mehfuz, Shabana; Kumar, Parmod
2016-01-01
Distribution system is the means of revenue for electric utility. It needs to be restored at the earliest if any feeder or complete system is tripped out due to fault or any other cause. Further, uncertainty of the loads, result in variations in the distribution network's parameters. Thus, an intelligent algorithm incorporating hybrid fuzzy-grey relation, which can take into account the uncertainties and compare the sequences is discussed to analyse and restore the distribution system. The simulation studies are carried out to show the utility of the method by ranking the restoration plans for a typical distribution system. This algorithm also meets the smart grid requirements in terms of an automated restoration plan for the partial/full blackout of network.
Shabetia, Alexander; Rodichev, Yurii; Veer, F.A.; Soroka, Elena; Louter, Christian; Bos, Freek; Belis, Jan; Veer, Fred; Nijsse, Rob
An analytical approach based on the on the sequential partitioning of the data and Weibull Statistical Distribution for inhomogeneous - defective materials is proposed. It allows assessing the guaranteed strength of glass structures for the low probability of fracture with a higher degree of
Energy Efficient Distributed Fault Identification Algorithm in Wireless Sensor Networks
Directory of Open Access Journals (Sweden)
Meenakshi Panda
2014-01-01
Full Text Available A distributed fault identification algorithm is proposed here to find both hard and soft faulty sensor nodes present in wireless sensor networks. The algorithm is distributed, self-detectable, and can detect the most common byzantine faults such as stuck at zero, stuck at one, and random data. In the proposed approach, each sensor node gathered the observed data from the neighbors and computed the mean to check whether faulty sensor node is present or not. If a node found the presence of faulty sensor node, then compares observed data with the data of the neighbors and predict probable fault status. The final fault status is determined by diffusing the fault information from the neighbors. The accuracy and completeness of the algorithm are verified with the help of statistical model of the sensors data. The performance is evaluated in terms of detection accuracy, false alarm rate, detection latency and message complexity.
Comparing Different Fault Identification Algorithms in Distributed Power System
Alkaabi, Salim
A power system is a huge complex system that delivers the electrical power from the generation units to the consumers. As the demand for electrical power increases, distributed power generation was introduced to the power system. Faults may occur in the power system at any time in different locations. These faults cause a huge damage to the system as they might lead to full failure of the power system. Using distributed generation in the power system made it even harder to identify the location of the faults in the system. The main objective of this work is to test the different fault location identification algorithms while tested on a power system with the different amount of power injected using distributed generators. As faults may lead the system to full failure, this is an important area for research. In this thesis different fault location identification algorithms have been tested and compared while the different amount of power is injected from distributed generators. The algorithms were tested on IEEE 34 node test feeder using MATLAB and the results were compared to find when these algorithms might fail and the reliability of these methods.
EDDA: An Efficient Distributed Data Replication Algorithm in VANETs.
Zhu, Junyu; Huang, Chuanhe; Fan, Xiying; Guo, Sipei; Fu, Bin
2018-02-10
Efficient data dissemination in vehicular ad hoc networks (VANETs) is a challenging issue due to the dynamic nature of the network. To improve the performance of data dissemination, we study distributed data replication algorithms in VANETs for exchanging information and computing in an arbitrarily-connected network of vehicle nodes. To achieve low dissemination delay and improve the network performance, we control the number of message copies that can be disseminated in the network and then propose an efficient distributed data replication algorithm (EDDA). The key idea is to let the data carrier distribute the data dissemination tasks to multiple nodes to speed up the dissemination process. We calculate the number of communication stages for the network to enter into a balanced status and show that the proposed distributed algorithm can converge to a consensus in a small number of communication stages. Most of the theoretical results described in this paper are to study the complexity of network convergence. The lower bound and upper bound are also provided in the analysis of the algorithm. Simulation results show that the proposed EDDA can efficiently disseminate messages to vehicles in a specific area with low dissemination delay and system overhead.
EDDA: An Efficient Distributed Data Replication Algorithm in VANETs
Directory of Open Access Journals (Sweden)
Junyu Zhu
2018-02-01
Full Text Available Efficient data dissemination in vehicular ad hoc networks (VANETs is a challenging issue due to the dynamic nature of the network. To improve the performance of data dissemination, we study distributed data replication algorithms in VANETs for exchanging information and computing in an arbitrarily-connected network of vehicle nodes. To achieve low dissemination delay and improve the network performance, we control the number of message copies that can be disseminated in the network and then propose an efficient distributed data replication algorithm (EDDA. The key idea is to let the data carrier distribute the data dissemination tasks to multiple nodes to speed up the dissemination process. We calculate the number of communication stages for the network to enter into a balanced status and show that the proposed distributed algorithm can converge to a consensus in a small number of communication stages. Most of the theoretical results described in this paper are to study the complexity of network convergence. The lower bound and upper bound are also provided in the analysis of the algorithm. Simulation results show that the proposed EDDA can efficiently disseminate messages to vehicles in a specific area with low dissemination delay and system overhead.
EDDA: An Efficient Distributed Data Replication Algorithm in VANETs
Zhu, Junyu; Huang, Chuanhe; Fan, Xiying; Guo, Sipei; Fu, Bin
2018-01-01
Efficient data dissemination in vehicular ad hoc networks (VANETs) is a challenging issue due to the dynamic nature of the network. To improve the performance of data dissemination, we study distributed data replication algorithms in VANETs for exchanging information and computing in an arbitrarily-connected network of vehicle nodes. To achieve low dissemination delay and improve the network performance, we control the number of message copies that can be disseminated in the network and then propose an efficient distributed data replication algorithm (EDDA). The key idea is to let the data carrier distribute the data dissemination tasks to multiple nodes to speed up the dissemination process. We calculate the number of communication stages for the network to enter into a balanced status and show that the proposed distributed algorithm can converge to a consensus in a small number of communication stages. Most of the theoretical results described in this paper are to study the complexity of network convergence. The lower bound and upper bound are also provided in the analysis of the algorithm. Simulation results show that the proposed EDDA can efficiently disseminate messages to vehicles in a specific area with low dissemination delay and system overhead. PMID:29439443
A Monte Carlo algorithm for the Vavilov distribution
International Nuclear Information System (INIS)
Yi, Chul-Young; Han, Hyon-Soo
1999-01-01
Using the convolution property of the inverse Laplace transform, an improved Monte Carlo algorithm for the Vavilov energy-loss straggling distribution of the charged particle is developed, which is relatively simple and gives enough accuracy to be used for most Monte Carlo applications
Applying Sequential Particle Swarm Optimization Algorithm to Improve Power Generation Quality
Directory of Open Access Journals (Sweden)
Abdulhafid Sallama
2014-10-01
Full Text Available Swarm Optimization approach is a heuristic search method whose mechanics are inspired by the swarming or collaborative behaviour of biological populations. It is used to solve constrained, unconstrained, continuous and discrete problems. Swarm intelligence systems are widely used and very effective in solving standard and large-scale optimization, provided that the problem does not require multi solutions. In this paper, particle swarm optimisation technique is used to optimise fuzzy logic controller (FLC for stabilising a power generation and distribution network that consists of four generators. The system is subject to different types of faults (single and multi-phase. Simulation studies show that the optimised FLC performs well in stabilising the network after it recovers from a fault. The controller is compared to multi-band and standard controllers.
A Degree Distribution Optimization Algorithm for Image Transmission
Jiang, Wei; Yang, Junjie
2016-09-01
Luby Transform (LT) code is the first practical implementation of digital fountain code. The coding behavior of LT code is mainly decided by the degree distribution which determines the relationship between source data and codewords. Two degree distributions are suggested by Luby. They work well in typical situations but not optimally in case of finite encoding symbols. In this work, the degree distribution optimization algorithm is proposed to explore the potential of LT code. Firstly selection scheme of sparse degrees for LT codes is introduced. Then probability distribution is optimized according to the selected degrees. In image transmission, bit stream is sensitive to the channel noise and even a single bit error may cause the loss of synchronization between the encoder and the decoder. Therefore the proposed algorithm is designed for image transmission situation. Moreover, optimal class partition is studied for image transmission with unequal error protection. The experimental results are quite promising. Compared with LT code with robust soliton distribution, the proposed algorithm improves the final quality of recovered images obviously with the same overhead.
A new distributed systems scheduling algorithm: a swarm intelligence approach
Haghi Kashani, Mostafa; Sarvizadeh, Raheleh; Jameii, Mahdi
2011-12-01
The scheduling problem in distributed systems is known as an NP-complete problem, and methods based on heuristic or metaheuristic search have been proposed to obtain optimal and suboptimal solutions. The task scheduling is a key factor for distributed systems to gain better performance. In this paper, an efficient method based on memetic algorithm is developed to solve the problem of distributed systems scheduling. With regard to load balancing efficiently, Artificial Bee Colony (ABC) has been applied as local search in the proposed memetic algorithm. The proposed method has been compared to existing memetic-Based approach in which Learning Automata method has been used as local search. The results demonstrated that the proposed method outperform the above mentioned method in terms of communication cost.
Principal distance constraint error diffusion algorithm for homogeneous dot distribution
Kang, Ki-Min; Kim, Choon-Woo
1999-12-01
The perceived quality of the halftoned image strongly depends on the spatial distribution of the binary dots. Various error diffusion algorithms have been proposed for realizing the homogeneous dot distribution in the highlight and shadow regions. However, they are computationally expensive and/or require large memory space. This paper presents a new threshold modulated error diffusion algorithm for the homogeneous dot distribution. The proposed method is applied exactly same as the Floyd-Steinberg's algorithm except the thresholding process. The threshold value is modulated based on the difference between the distance to the nearest minor pixel, `minor pixel distance', and the principal distance. To do so, calculation of the minor pixel distance is needed for every pixel. But, it is quite time consuming and requires large memory resources. In order to alleviate this problem, `the minor pixel offset array' that transforms the 2D history of minor pixels into the 1D codes is proposed. The proposed algorithm drastically reduces the computational load and memory spaces needed for calculation of the minor pixel distance.
Online algorithms for optimal energy distribution in microgrids
Wang, Yu; Nelms, R Mark
2015-01-01
Presenting an optimal energy distribution strategy for microgrids in a smart grid environment, and featuring a detailed analysis of the mathematical techniques of convex optimization and online algorithms, this book provides readers with essential content on how to achieve multi-objective optimization that takes into consideration power subscribers, energy providers and grid smoothing in microgrids. Featuring detailed theoretical proofs and simulation results that demonstrate and evaluate the correctness and effectiveness of the algorithm, this text explains step-by-step how the problem can b
Li, Jinze; Qu, Zhi; He, Xiaoyang; Jin, Xiaoming; Li, Tie; Wang, Mingkai; Han, Qiu; Gao, Ziji; Jiang, Feng
2018-02-01
Large-scale access of distributed power can improve the current environmental pressure, at the same time, increasing the complexity and uncertainty of overall distribution system. Rational planning of distributed power can effectively improve the system voltage level. To this point, the specific impact on distribution network power quality caused by the access of typical distributed power was analyzed and from the point of improving the learning factor and the inertia weight, an improved particle swarm optimization algorithm (IPSO) was proposed which could solve distributed generation planning for distribution network to improve the local and global search performance of the algorithm. Results show that the proposed method can well reduce the system network loss and improve the economic performance of system operation with distributed generation.
Research reactor loading pattern optimization using estimation of distribution algorithms
Energy Technology Data Exchange (ETDEWEB)
Jiang, S. [Dept. of Earth Science and Engineering, Applied Modeling and Computation Group AMCG, Imperial College, London, SW7 2AZ (United Kingdom); Ziver, K. [Dept. of Earth Science and Engineering, Applied Modeling and Computation Group AMCG, Imperial College, London, SW7 2AZ (United Kingdom); AMCG Group, RM Consultants, Abingdon (United Kingdom); Carter, J. N.; Pain, C. C.; Eaton, M. D.; Goddard, A. J. H. [Dept. of Earth Science and Engineering, Applied Modeling and Computation Group AMCG, Imperial College, London, SW7 2AZ (United Kingdom); Franklin, S. J.; Phillips, H. J. [Imperial College, Reactor Centre, Silwood Park, Buckhurst Road, Ascot, Berkshire, SL5 7TE (United Kingdom)
2006-07-01
A new evolutionary search based approach for solving the nuclear reactor loading pattern optimization problems is presented based on the Estimation of Distribution Algorithms. The optimization technique developed is then applied to the maximization of the effective multiplication factor (K{sub eff}) of the Imperial College CONSORT research reactor (the last remaining civilian research reactor in the United Kingdom). A new elitism-guided searching strategy has been developed and applied to improve the local convergence together with some problem-dependent information based on the 'stand-alone K{sub eff} with fuel coupling calculations. A comparison study between the EDAs and a Genetic Algorithm with Heuristic Tie Breaking Crossover operator has shown that the new algorithm is efficient and robust. (authors)
Research reactor loading pattern optimization using estimation of distribution algorithms
International Nuclear Information System (INIS)
Jiang, S.; Ziver, K.; Carter, J. N.; Pain, C. C.; Eaton, M. D.; Goddard, A. J. H.; Franklin, S. J.; Phillips, H. J.
2006-01-01
A new evolutionary search based approach for solving the nuclear reactor loading pattern optimization problems is presented based on the Estimation of Distribution Algorithms. The optimization technique developed is then applied to the maximization of the effective multiplication factor (K eff ) of the Imperial College CONSORT research reactor (the last remaining civilian research reactor in the United Kingdom). A new elitism-guided searching strategy has been developed and applied to improve the local convergence together with some problem-dependent information based on the 'stand-alone K eff with fuel coupling calculations. A comparison study between the EDAs and a Genetic Algorithm with Heuristic Tie Breaking Crossover operator has shown that the new algorithm is efficient and robust. (authors)
International Nuclear Information System (INIS)
Botet, R.
1996-01-01
A new kinetic fragmentation model, the Fragmentation - Inactivation -Binary (FIB) model is described where a dissipative process stops randomly the sequential, conservative and off-equilibrium fragmentation process. (K.A.)
Distribution Agnostic Structured Sparsity Recovery: Algorithms and Applications
Masood, Mudassir
2015-05-01
Compressed sensing has been a very active area of research and several elegant algorithms have been developed for the recovery of sparse signals in the past few years. However, most of these algorithms are either computationally expensive or make some assumptions that are not suitable for all real world problems. Recently, focus has shifted to Bayesian-based approaches that are able to perform sparse signal recovery at much lower complexity while invoking constraint and/or a priori information about the data. While Bayesian approaches have their advantages, these methods must have access to a priori statistics. Usually, these statistics are unknown and are often difficult or even impossible to predict. An effective workaround is to assume a distribution which is typically considered to be Gaussian, as it makes many signal processing problems mathematically tractable. Seemingly attractive, this assumption necessitates the estimation of the associated parameters; which could be hard if not impossible. In the thesis, we focus on this aspect of Bayesian recovery and present a framework to address the challenges mentioned above. The proposed framework allows Bayesian recovery of sparse signals but at the same time is agnostic to the distribution of the unknown sparse signal components. The algorithms based on this framework are agnostic to signal statistics and utilize a priori statistics of additive noise and the sparsity rate of the signal, which are shown to be easily estimated from data if not available. In the thesis, we propose several algorithms based on this framework which utilize the structure present in signals for improved recovery. In addition to the algorithm that considers just the sparsity structure of sparse signals, tools that target additional structure of the sparsity recovery problem are proposed. These include several algorithms for a) block-sparse signal estimation, b) joint reconstruction of several common support sparse signals, and c
Directory of Open Access Journals (Sweden)
Tingsong Du
2015-01-01
Full Text Available An improved quantum artificial fish swarm algorithm (IQAFSA for solving distributed network programming considering distributed generation is proposed in this work. The IQAFSA based on quantum computing which has exponential acceleration for heuristic algorithm uses quantum bits to code artificial fish and quantum revolving gate, preying behavior, and following behavior and variation of quantum artificial fish to update the artificial fish for searching for optimal value. Then, we apply the proposed new algorithm, the quantum artificial fish swarm algorithm (QAFSA, the basic artificial fish swarm algorithm (BAFSA, and the global edition artificial fish swarm algorithm (GAFSA to the simulation experiments for some typical test functions, respectively. The simulation results demonstrate that the proposed algorithm can escape from the local extremum effectively and has higher convergence speed and better accuracy. Finally, applying IQAFSA to distributed network problems and the simulation results for 33-bus radial distribution network system show that IQAFSA can get the minimum power loss after comparing with BAFSA, GAFSA, and QAFSA.
Improving permafrost distribution modelling using feature selection algorithms
Deluigi, Nicola; Lambiel, Christophe; Kanevski, Mikhail
2016-04-01
The availability of an increasing number of spatial data on the occurrence of mountain permafrost allows the employment of machine learning (ML) classification algorithms for modelling the distribution of the phenomenon. One of the major problems when dealing with high-dimensional dataset is the number of input features (variables) involved. Application of ML classification algorithms to this large number of variables leads to the risk of overfitting, with the consequence of a poor generalization/prediction. For this reason, applying feature selection (FS) techniques helps simplifying the amount of factors required and improves the knowledge on adopted features and their relation with the studied phenomenon. Moreover, taking away irrelevant or redundant variables from the dataset effectively improves the quality of the ML prediction. This research deals with a comparative analysis of permafrost distribution models supported by FS variable importance assessment. The input dataset (dimension = 20-25, 10 m spatial resolution) was constructed using landcover maps, climate data and DEM derived variables (altitude, aspect, slope, terrain curvature, solar radiation, etc.). It was completed with permafrost evidences (geophysical and thermal data and rock glacier inventories) that serve as training permafrost data. Used FS algorithms informed about variables that appeared less statistically important for permafrost presence/absence. Three different algorithms were compared: Information Gain (IG), Correlation-based Feature Selection (CFS) and Random Forest (RF). IG is a filter technique that evaluates the worth of a predictor by measuring the information gain with respect to the permafrost presence/absence. Conversely, CFS is a wrapper technique that evaluates the worth of a subset of predictors by considering the individual predictive ability of each variable along with the degree of redundancy between them. Finally, RF is a ML algorithm that performs FS as part of its
ZAP: a distributed channel assignment algorithm for cognitive radio networks
Junior , Paulo Roberto ,; Fonseca , Mauro; Munaretto , Anelise; Viana , Aline ,; Ziviani , Artur
2011-01-01
Abstract We propose ZAP, an algorithm for the distributed channel assignment in cognitive radio (CR) networks. CRs are capable of identifying underutilized licensed bands of the spectrum, allowing their reuse by secondary users without interfering with primary users. In this context, efficient channel assignment is challenging as ideally it must be simple, incur acceptable communication overhead, provide timely response, and be adaptive to accommodate frequent changes in the network. Another ...
A Distributed Algorithm for Energy Optimization in Hydraulic Networks
DEFF Research Database (Denmark)
Kallesøe, Carsten; Wisniewski, Rafal; Jensen, Tom Nørgaard
2014-01-01
An industrial case study in the form of a large-scale hydraulic network underlying a district heating system is considered. A distributed control is developed that minimizes the aggregated electrical energy consumption of the pumps in the network without violating the control demands. The algorithm...... a Plug & Play control system as most commissioning can be done during the manufacture of the pumps. Only information on the graph-structure of the hydraulic network is needed during installation....
Distributed Scheduling in Time Dependent Environments: Algorithms and Analysis
Shmuel, Ori; Cohen, Asaf; Gurewitz, Omer
2017-01-01
Consider the problem of a multiple access channel in a time dependent environment with a large number of users. In such a system, mostly due to practical constraints (e.g., decoding complexity), not all users can be scheduled together, and usually only one user may transmit at any given time. Assuming a distributed, opportunistic scheduling algorithm, we analyse the system's properties, such as delay, QoS and capacity scaling laws. Specifically, we start with analyzing the performance while \\...
Françoise Benz
2004-01-01
ACADEMIC TRAINING LECTURE REGULAR PROGRAMME 1, 2, 3 and 4 June From 11:00 hrs to 12:00 hrs - Main Auditorium bldg. 500 Evolutionary Heuristic Optimization: Genetic Algorithms and Estimation of Distribution Algorithms V. Robles Forcada and M. Perez Hernandez / Univ. de Madrid, Spain In the real world, there exist a huge number of problems that require getting an optimum or near-to-optimum solution. Optimization can be used to solve a lot of different problems such as network design, sets and partitions, storage and retrieval or scheduling. On the other hand, in nature, there exist many processes that seek a stable state. These processes can be seen as natural optimization processes. Over the last 30 years several attempts have been made to develop optimization algorithms, which simulate these natural optimization processes. These attempts have resulted in methods such as Simulated Annealing, based on natural annealing processes or Evolutionary Computation, based on biological evolution processes. Geneti...
Françoise Benz
2004-01-01
ENSEIGNEMENT ACADEMIQUE ACADEMIC TRAINING Françoise Benz 73127 academic.training@cern.ch ACADEMIC TRAINING LECTURE REGULAR PROGRAMME 1, 2, 3 and 4 June From 11:00 hrs to 12:00 hrs - Main Auditorium bldg. 500 Evolutionary Heuristic Optimization: Genetic Algorithms and Estimation of Distribution Algorithms V. Robles Forcada and M. Perez Hernandez / Univ. de Madrid, Spain In the real world, there exist a huge number of problems that require getting an optimum or near-to-optimum solution. Optimization can be used to solve a lot of different problems such as network design, sets and partitions, storage and retrieval or scheduling. On the other hand, in nature, there exist many processes that seek a stable state. These processes can be seen as natural optimization processes. Over the last 30 years several attempts have been made to develop optimization algorithms, which simulate these natural optimization processes. These attempts have resulted in methods such as Simulated Annealing, based on nat...
Qin, Jiahu; Fu, Weiming; Gao, Huijun; Zheng, Wei Xing
2016-03-03
This paper is concerned with developing a distributed k-means algorithm and a distributed fuzzy c-means algorithm for wireless sensor networks (WSNs) where each node is equipped with sensors. The underlying topology of the WSN is supposed to be strongly connected. The consensus algorithm in multiagent consensus theory is utilized to exchange the measurement information of the sensors in WSN. To obtain a faster convergence speed as well as a higher possibility of having the global optimum, a distributed k-means++ algorithm is first proposed to find the initial centroids before executing the distributed k-means algorithm and the distributed fuzzy c-means algorithm. The proposed distributed k-means algorithm is capable of partitioning the data observed by the nodes into measure-dependent groups which have small in-group and large out-group distances, while the proposed distributed fuzzy c-means algorithm is capable of partitioning the data observed by the nodes into different measure-dependent groups with degrees of membership values ranging from 0 to 1. Simulation results show that the proposed distributed algorithms can achieve almost the same results as that given by the centralized clustering algorithms.
Parallel SN algorithms in shared- and distributed-memory environments
International Nuclear Information System (INIS)
Haghighat, Alireza; Hunter, Melissa A.; Mattis, Ronald E.
1995-01-01
Different 2-D spatial domain partitioning Sn transport theory algorithms have been developed on the basis of the Block-Jacobi iterative scheme. These algorithms have been incorporated into TWOTRAN-II, and tested on a shared-memory CRAY Y-MP C90 and a distributed-memory IBM SP1. For a series of fixed source r-z geometry homogeneous problems, parallel efficiencies in a range of 50-90% are achieved on the C90 with 6 processors, and lower values (20-60%) are obtained on the SP1. It is demonstrated that better performance is attainable if one addresses issues such as convergence rate, load-balancing, and granularity for both architectures, as well as message passing (network bandwidth and latency) for SP1. (author). 17 refs, 4 figs
ZAP: a distributed channel assignment algorithm for cognitive radio networks
Directory of Open Access Journals (Sweden)
Munaretto Anelise
2011-01-01
Full Text Available Abstract We propose ZAP, an algorithm for the distributed channel assignment in cognitive radio (CR networks. CRs are capable of identifying underutilized licensed bands of the spectrum, allowing their reuse by secondary users without interfering with primary users. In this context, efficient channel assignment is challenging as ideally it must be simple, incur acceptable communication overhead, provide timely response, and be adaptive to accommodate frequent changes in the network. Another challenge is the optimization of network capacity through interference minimization. In contrast to related work, ZAP addresses these challenges with a fully distributed approach based only on local (neighborhood knowledge, while significantly reducing computational costs and the number of messages required for channel assignment. Simulations confirm the efficiency of ZAP in terms of (i the performance tradeoff between different metrics and (ii the fast achievement of a suitable assignment solution regardless of network size and density.
Combustion distribution control using the extremum seeking algorithm
Marjanovic, A.; Krstic, M.; Djurovic, Z.; Kvascev, G.; Papic, V.
2014-12-01
Quality regulation of the combustion process inside the furnace is the basis of high demands for increasing robustness, safety and efficiency of thermal power plants. The paper considers the possibility of spatial temperature distribution control inside the boiler, based on the correction of distribution of coal over the mills. Such control system ensures the maintenance of the flame focus away from the walls of the boiler, and thus preserves the equipment and reduces the possibility of ash slugging. At the same time, uniform heat dissipation over mills enhances the energy efficiency of the boiler, while reducing the pollution of the system. A constrained multivariable extremum seeking algorithm is proposed as a tool for combustion process optimization with the main objective of centralizing the flame in the furnace. Simulations are conducted on a model corresponding to the 350MW boiler of the Nikola Tesla Power Plant, in Obrenovac, Serbia.
Combustion distribution control using the extremum seeking algorithm
International Nuclear Information System (INIS)
Marjanovic, A; Djurovic, Z; Kvascev, G; Papic, V; Krstic, M
2014-01-01
Quality regulation of the combustion process inside the furnace is the basis of high demands for increasing robustness, safety and efficiency of thermal power plants. The paper considers the possibility of spatial temperature distribution control inside the boiler, based on the correction of distribution of coal over the mills. Such control system ensures the maintenance of the flame focus away from the walls of the boiler, and thus preserves the equipment and reduces the possibility of ash slugging. At the same time, uniform heat dissipation over mills enhances the energy efficiency of the boiler, while reducing the pollution of the system. A constrained multivariable extremum seeking algorithm is proposed as a tool for combustion process optimization with the main objective of centralizing the flame in the furnace. Simulations are conducted on a model corresponding to the 350MW boiler of the Nikola Tesla Power Plant, in Obrenovac, Serbia
DEFF Research Database (Denmark)
Hansen, Thomas Mejer; Mosegaard, Klaus; Cordua, Knud Skou
2010-01-01
Markov chain Monte Carlo methods such as the Gibbs sampler and the Metropolis algorithm can be used to sample the solutions to non-linear inverse problems. In principle these methods allow incorporation of arbitrarily complex a priori information, but current methods allow only relatively simple...... this algorithm with the Metropolis algorithm to obtain an efficient method for sampling posterior probability densities for nonlinear inverse problems....
Parallel discrete ordinates algorithms on distributed and common memory systems
International Nuclear Information System (INIS)
Wienke, B.R.; Hiromoto, R.E.; Brickner, R.G.
1987-01-01
The S/sub n/ algorithm employs iterative techniques in solving the linear Boltzmann equation. These methods, both ordered and chaotic, were compared on both the Denelcor HEP and the Intel hypercube. Strategies are linked to the organization and accessibility of memory (common memory versus distributed memory architectures), with common concern for acquisition of global information. Apart from this, the inherent parallelism of the algorithm maps directly onto the two architectures. Results comparing execution times, speedup, and efficiency are based on a representative 16-group (full upscatter and downscatter) sample problem. Calculations were performed on both the Los Alamos National Laboratory (LANL) Denelcor HEP and the LANL Intel hypercube. The Denelcor HEP is a 64-bit multi-instruction, multidate MIMD machine consisting of up to 16 process execution modules (PEMs), each capable of executing 64 processes concurrently. Each PEM can cooperate on a job, or run several unrelated jobs, and share a common global memory through a crossbar switch. The Intel hypercube, on the other hand, is a distributed memory system composed of 128 processing elements, each with its own local memory. Processing elements are connected in a nearest-neighbor hypercube configuration and sharing of data among processors requires execution of explicit message-passing constructs
Efficient distribution of toy products using ant colony optimization algorithm
Hidayat, S.; Nurpraja, C. A.
2017-12-01
CV Atham Toys (CVAT) produces wooden toys and furniture, comprises 13 small and medium industries. CVAT always attempt to deliver customer orders on time but delivery costs are high. This is because of inadequate infrastructure such that delivery routes are long, car maintenance costs are high, while fuel subsidy by the government is still temporary. This study seeks to minimize the cost of product distribution based on the shortest route using one of five Ant Colony Optimization (ACO) algorithms to solve the Vehicle Routing Problem (VRP). This study concludes that the best of the five is the Ant Colony System (ACS) algorithm. The best route in 1st week gave a total distance of 124.11 km at a cost of Rp 66,703.75. The 2nd week route gave a total distance of 132.27 km at a cost of Rp 71,095.13. The 3rd week best route gave a total distance of 122.70 km with a cost of Rp 65,951.25. While the 4th week gave a total distance of 132.27 km at a cost of Rp 74,083.63. Prior to this study there was no effort to calculate these figures.
Fast Parabola Detection Using Estimation of Distribution Algorithms
Directory of Open Access Journals (Sweden)
Jose de Jesus Guerrero-Turrubiates
2017-01-01
Full Text Available This paper presents a new method based on Estimation of Distribution Algorithms (EDAs to detect parabolic shapes in synthetic and medical images. The method computes a virtual parabola using three random boundary pixels to calculate the constant values of the generic parabola equation. The resulting parabola is evaluated by matching it with the parabolic shape in the input image by using the Hadamard product as fitness function. This proposed method is evaluated in terms of computational time and compared with two implementations of the generalized Hough transform and RANSAC method for parabola detection. Experimental results show that the proposed method outperforms the comparative methods in terms of execution time about 93.61% on synthetic images and 89% on retinal fundus and human plantar arch images. In addition, experimental results have also shown that the proposed method can be highly suitable for different medical applications.
A modified estimation distribution algorithm based on extreme elitism.
Gao, Shujun; de Silva, Clarence W
2016-12-01
An existing estimation distribution algorithm (EDA) with univariate marginal Gaussian model was improved by designing and incorporating an extreme elitism selection method. This selection method highlighted the effect of a few top best solutions in the evolution and advanced EDA to form a primary evolution direction and obtain a fast convergence rate. Simultaneously, this selection can also keep the population diversity to make EDA avoid premature convergence. Then the modified EDA was tested by means of benchmark low-dimensional and high-dimensional optimization problems to illustrate the gains in using this extreme elitism selection. Besides, no-free-lunch theorem was implemented in the analysis of the effect of this new selection on EDAs. Copyright Â© 2016 Elsevier Ireland Ltd. All rights reserved.
Directory of Open Access Journals (Sweden)
Altemeier William A
2010-05-01
Full Text Available Abstract Background Ventilator-induced lung injury (VILI is a recognized complication of mechanical ventilation. Although the specific mechanism by which mechanical ventilation causes lung injury remains an active area of study, the application of positive end expiratory pressure (PEEP reduces its severity. We have previously reported that VILI is spatially heterogeneous with the most severe injury in the dorsal-caudal lung. This regional injury heterogeneity was abolished by the application of PEEP = 8 cm H2O. We hypothesized that the spatial distribution of lung injury correlates with areas in which cyclical airway collapse and recruitment occurs. Methods To test this hypothesis, rabbits were mechanically ventilated in the supine posture, and regional ventilation distribution was measured under four conditions: tidal volumes (VT of 6 and 12 ml/kg with PEEP levels of 0 and 8 cm H2O. Results We found that relative ventilation was sequentially redistributed towards dorsal-caudal lung with increasing tidal volume. This sequential ventilation redistribution was abolished with the addition of PEEP. Conclusions These results suggest that cyclical airway collapse and recruitment is regionally heterogeneous and spatially correlated with areas most susceptible to VILI.
Distributed data access in the sequential access model at the D0 experiment at Fermilab
International Nuclear Information System (INIS)
Terekhov, Igor; White, Victoria
2000-01-01
The authors present the Sequential Access Model (SAM), which is the data handling system for D0, one of two primary High Energy Experiments at Fermilab. During the next several years, the D0 experiment will store a total of about 1 PByte of data, including raw detector data and data processed at various levels. The design of SAM is not specific to the D0 experiment and carries few assumptions about the underlying mass storage level; its ideas are applicable to any sequential data access. By definition, in the sequential access mode a user application needs to process a stream of data, by accessing each data unit exactly once, the order of data units in the stream being irrelevant. The units of data are laid out sequentially in files. The adopted model allows for significant optimizations of system performance, decrease of user file latency and increase of overall throughput. In particular, caching is done with the knowledge of all the files needed in the near future, defined as all the files of the already running or submitted jobs. The bulk of the data is stored in files on tape in the mass storage system (MSS) called Enstore[2] and also developed at Fermilab. (The tape drives are served by an ADIC AML/2 Automated Tape Library). At any given time, SAM has a small fraction of the data cached on disk for processing. In the present paper, the authors discuss how data is delivered onto disk and how it is accessed by user applications. They will concentrate on data retrieval (consumption) from the MSS; when SAM is used for storing of data, the mechanisms are rather symmetrical. All of the data managed by SAM is cataloged in great detail in a relational database (ORACLE). The database also serves as the persistency mechanism for the SAM servers described in this paper. Any client or server in the SAM system which needs to store or retrieve information from the database does so through the interfaces of a CORBA-based database server. The users (physicists) use the
The derivation of distributed termination detection algorithms from garbage collection schemes
Tel, G.; Mattern, F.
1990-01-01
It is shown that the termination detection problem for distributed computations can be modelled as an instance of the garbage collection problem. Consequently, algorithms for the termination detection problem are obtained by applying transformations to garbage collection algorithms. The
Fuzzy Dynamic Discrimination Algorithms for Distributed Knowledge Management Systems
Directory of Open Access Journals (Sweden)
Vasile MAZILESCU
2010-12-01
Full Text Available A reduction of the algorithmic complexity of the fuzzy inference engine has the following property: the inputs (the fuzzy rules and the fuzzy facts can be divided in two parts, one being relatively constant for a long a time (the fuzzy rule or the knowledge model when it is compared to the second part (the fuzzy facts for every inference cycle. The occurrence of certain transformations over the constant part makes sense, in order to decrease the solution procurement time, in the case that the second part varies, but it is known at certain moments in time. The transformations attained in advance are called pre-processing or knowledge compilation. The use of variables in a Business Rule Management System knowledge representation allows factorising knowledge, like in classical knowledge based systems. The language of the first-degree predicates facilitates the formulation of complex knowledge in a rigorous way, imposing appropriate reasoning techniques. It is, thus, necessary to define the description method of fuzzy knowledge, to justify the knowledge exploiting efficiency when the compiling technique is used, to present the inference engine and highlight the functional features of the pattern matching and the state space processes. This paper presents the main results of our project PR356 for designing a compiler for fuzzy knowledge, like Rete compiler, that comprises two main components: a static fuzzy discrimination structure (Fuzzy Unification Tree and the Fuzzy Variables Linking Network. There are also presented the features of the elementary pattern matching process that is based on the compiled structure of fuzzy knowledge. We developed fuzzy discrimination algorithms for Distributed Knowledge Management Systems (DKMSs. The implementations have been elaborated in a prototype system FRCOM (Fuzzy Rule COMpiler.
Jing Chen
2015-01-01
This study takes the concept of food logistics distribution as the breakthrough point, by means of the aim of optimization of food logistics distribution routes and analysis of the optimization model of food logistics route, as well as the interpretation of the genetic algorithm, it discusses the optimization of food logistics distribution route based on genetic and cluster scheme algorithm.
Liu, Hua-Long; Liu, Hua-Dong
2014-10-01
Partial discharge (PD) in power transformers is one of the prime reasons resulting in insulation degradation and power faults. Hence, it is of great importance to study the techniques of the detection and localization of PD in theory and practice. The detection and localization of PD employing acoustic emission (AE) techniques, as a kind of non-destructive testing, plus due to the advantages of powerful capability of locating and high precision, have been paid more and more attention. The localization algorithm is the key factor to decide the localization accuracy in AE localization of PD. Many kinds of localization algorithms exist for the PD source localization adopting AE techniques including intelligent and non-intelligent algorithms. However, the existed algorithms possess some defects such as the premature convergence phenomenon, poor local optimization ability and unsuitability for the field applications. To overcome the poor local optimization ability and easily caused premature convergence phenomenon of the fundamental genetic algorithm (GA), a new kind of improved GA is proposed, namely the sequence quadratic programming-genetic algorithm (SQP-GA). For the hybrid optimization algorithm, SQP-GA, the sequence quadratic programming (SQP) algorithm which is used as a basic operator is integrated into the fundamental GA, so the local searching ability of the fundamental GA is improved effectively and the premature convergence phenomenon is overcome. Experimental results of the numerical simulations of benchmark functions show that the hybrid optimization algorithm, SQP-GA, is better than the fundamental GA in the convergence speed and optimization precision, and the proposed algorithm in this paper has outstanding optimization effect. At the same time, the presented SQP-GA in the paper is applied to solve the ultrasonic localization problem of PD in transformers, then the ultrasonic localization method of PD in transformers based on the SQP-GA is proposed. And
Performance Analysis of the Consensus-Based Distributed LMS Algorithm
Directory of Open Access Journals (Sweden)
Gonzalo Mateos
2009-01-01
Full Text Available Low-cost estimation of stationary signals and reduced-complexity tracking of nonstationary processes are well motivated tasks than can be accomplished using ad hoc wireless sensor networks (WSNs. To this end, a fully distributed least mean-square (D-LMS algorithm is developed in this paper, in which sensors exchange messages with single-hop neighbors to consent on the network-wide estimates adaptively. The novel approach does not require a Hamiltonian cycle or a special bridge subset of sensors, while communications among sensors are allowed to be noisy. A mean-square error (MSE performance analysis of D-LMS is conducted in the presence of a time-varying parameter vector, which adheres to a first-order autoregressive model. For sensor observations that are related to the parameter vector of interest via a linear Gaussian model and after adopting simplifying independence assumptions, exact closed-form expressions are derived for the global and sensor-level MSE evolution as well as its steady-state (s.s. values. Mean and MSE-sense stability of D-LMS are also established. Interestingly, extensive numerical tests demonstrate that for small step-sizes the results accurately extend to the pragmatic setting whereby sensors acquire temporally correlated, not necessarily Gaussian data.
Monte Carlo dose calculation algorithm on a distributed system
International Nuclear Information System (INIS)
Chauvie, Stephane; Dominoni, Matteo; Marini, Piergiorgio; Stasi, Michele; Pia, Maria Grazia; Scielzo, Giuseppe
2003-01-01
The main goal of modern radiotherapy, such as 3D conformal radiotherapy and intensity-modulated radiotherapy is to deliver a high dose to the target volume sparing the surrounding healthy tissue. The accuracy of dose calculation in a treatment planning system is therefore a critical issue. Among many algorithms developed over the last years, those based on Monte Carlo proven to be very promising in terms of accuracy. The most severe obstacle in application to clinical practice is the high time necessary for calculations. We have studied a high performance network of Personal Computer as a realistic alternative to a high-costs dedicated parallel hardware to be used routinely as instruments of evaluation of treatment plans. We set-up a Beowulf Cluster, configured with 4 nodes connected with low-cost network and installed MC code Geant4 to describe our irradiation facility. The MC, once parallelised, was run on the Beowulf Cluster. The first run of the full simulation showed that the time required for calculation decreased linearly increasing the number of distributed processes. The good scalability trend allows both statistically significant accuracy and good time performances. The scalability of the Beowulf Cluster system offers a new instrument for dose calculation that could be applied in clinical practice. These would be a good support particularly in high challenging prescription that needs good calculation accuracy in zones of high dose gradient and great dishomogeneities
Directory of Open Access Journals (Sweden)
Tine L. Vandoorn
2015-06-01
Full Text Available The increasing share of distributed energy resources poses a challenge to the distribution network operator (DNO to maintain the current availability of the system while limiting the investment costs. Related to this, there is a clear trend in DNOs trying to better monitor their grid by installing a distribution management system (DMS. This DMS enables the DNOs to remotely switch their network or better localize and solve faults. Moreover, the DMS can be used to centrally control the grid assets. Therefore, in this paper, a control strategy is discussed that can be implemented in the DMS for solving current congestion problems posed by the increasing share of renewables in the grid. This control strategy controls wind turbines in order to avoid congestion while mitigating the required investment costs in order to achieve a global cost-efficient solution. Next to the application and objective of the control, the parameter tuning of the control algorithm is discussed.
Low-Complexity Compression Algorithm for Hyperspectral Images Based on Distributed Source Coding
Directory of Open Access Journals (Sweden)
Yongjian Nian
2013-01-01
Full Text Available A low-complexity compression algorithm for hyperspectral images based on distributed source coding (DSC is proposed in this paper. The proposed distributed compression algorithm can realize both lossless and lossy compression, which is implemented by performing scalar quantization strategy on the original hyperspectral images followed by distributed lossless compression. Multilinear regression model is introduced for distributed lossless compression in order to improve the quality of side information. Optimal quantized step is determined according to the restriction of the correct DSC decoding, which makes the proposed algorithm achieve near lossless compression. Moreover, an effective rate distortion algorithm is introduced for the proposed algorithm to achieve low bit rate. Experimental results show that the compression performance of the proposed algorithm is competitive with that of the state-of-the-art compression algorithms for hyperspectral images.
Study of the distribution of maxima and minima in multiple sequential images of uniformity
International Nuclear Information System (INIS)
Llacer Martos, S.; Puchal Ane, R.
2011-01-01
To characterize the uniformity of a gamma camera extrinsic used integral uniformity coefficient is calculated with the value of two pixels, the maximum and minimum, single source acquisition of a flat and uniform. This method does not take into account the fact that if a gamma camera having a uniform response, the distribution of these items should be random. In this paper we study how these points are distributed in a succession of large numbers of uniform images.
International Nuclear Information System (INIS)
Taniguchi, Yuu; Okano, Shigetaka; Mochizuki, Masahito
2017-01-01
The excessive tensile residual stress generated by welding after surface machining may be an important factor to cause stress corrosion cracking (SCC) in nuclear power plants. Therefore we need to understand and control the residual stress distribution appropriately. In this study, residual stress distributions within surface machined layer generated by surface machining and sequential welding were evaluated by X-ray diffraction method. Depth directional distributions were also investigated by electrolytic polishing. In addition, to consider the effect of work hardened layer on the residual stress distributions, we also measured full width at half maximum (FWHM) obtained from X-ray diffraction. Testing material was a low-carbon austenitic stainless steel type SUS316L. Test specimens were prepared by surface machining with different cutting conditions. Then, bead-on-plate welding under the same welding condition was carried out on the test specimens with different surface machined layer. As a result, the tensile residual stress generated by surface machining increased with increasing cutting speed and showed nearly uniform distributions on the surface. Furthermore, the tensile residual stress drastically decreased with increasing measurement depth within surface machined layer. Then, the residual stress approached 0 MPa after the compressive value showed. FWHM also decreased drastically with increasing measurement depth and almost constant value from a certain depth, which was almost equal regardless of the machining condition, within surface machined layer in all specimens. After welding, the transverse distribution of the longitudinal residual stress varied in the area apart from the weld center according to machining conditions and had a maximum value in heat affected zone. The magnitude of the maximum residual stress was almost equal regardless of the machining condition and decreased with increasing measurement depth within surface machined layer. Finally, the
International Nuclear Information System (INIS)
Niknam, Taher; Azadfarsani, Ehsan; Jabbari, Masoud
2012-01-01
Highlights: ► Network reconfiguration is a very important way to save the electrical energy. ► This paper proposes a new algorithm to solve the DFR. ► The algorithm combines NFAPSO with NM. ► The proposed algorithm is tested on two distribution test feeders. - Abstract: Network reconfiguration for loss reduction in distribution system is a very important way to save the electrical energy. This paper proposes a new hybrid evolutionary algorithm to solve the Distribution Feeder Reconfiguration problem (DFR). The algorithm is based on combination of a New Fuzzy Adaptive Particle Swarm Optimization (NFAPSO) and Nelder–Mead simplex search method (NM) called NFAPSO–NM. In the proposed algorithm, a new fuzzy adaptive particle swarm optimization includes two parts. The first part is Fuzzy Adaptive Binary Particle Swarm Optimization (FABPSO) that determines the status of tie switches (open or close) and second part is Fuzzy Adaptive Discrete Particle Swarm Optimization (FADPSO) that determines the sectionalizing switch number. In other side, due to the results of binary PSO(BPSO) and discrete PSO(DPSO) algorithms highly depends on the values of their parameters such as the inertia weight and learning factors, a fuzzy system is employed to adaptively adjust the parameters during the search process. Moreover, the Nelder–Mead simplex search method is combined with the NFAPSO algorithm to improve its performance. Finally, the proposed algorithm is tested on two distribution test feeders. The results of simulation show that the proposed method is very powerful and guarantees to obtain the global optimization.
Directory of Open Access Journals (Sweden)
Niti Ashish Kumar Desai
2015-12-01
Full Text Available Business Strategies are formulated based on an understanding of customer needs. This requires development of a strategy to understand customer behaviour and buying patterns, both current and future. This involves understanding, first how an organization currently understands customer needs and second predicting future trends to drive growth. This article focuses on purchase trend of customer, where timing of purchase is more important than association of item to be purchased, and which can be found out with Sequential Pattern Mining (SPM methods. Conventional SPM algorithms worked purely on frequency identifying patterns that were more frequent but suffering from challenges like generation of huge number of uninteresting patterns, lack of user’s interested patterns, rare item problem, etc. Article attempts a solution through development of a SPM algorithm based on various constraints like Gap, Compactness, Item, Recency, Profitability and Length along with Frequency constraint. Incorporation of six additional constraints is as well to ensure that all patterns are recently active (Recency, active for certain time span (Compactness, profitable and indicative of next timeline for purchase (Length―Item―Gap. The article also attempts to throw light on how proposed Constraint-based Prefix Span algorithm is helpful to understand buying behaviour of customer which is in formative stage.
Dynamic Consensus Algorithm based Distributed Voltage Harmonic Compensation in Islanded Microgrids
DEFF Research Database (Denmark)
Meng, Lexuan; Tang, Fen; Firoozabadi, Mehdi Savaghebi
2015-01-01
generators can be employed as compensators to enhance the power quality on consumer side. However, conventional centralized control is facing obstacles because of the distributed fashion of generation and consumption. Accordingly, this paper proposes a consensus algorithm based distributed hierarchical...
Hybrid fuzzy charged system search algorithm based state estimation in distribution networks
Directory of Open Access Journals (Sweden)
Sachidananda Prasad
2017-06-01
Full Text Available This paper proposes a new hybrid charged system search (CSS algorithm based state estimation in radial distribution networks in fuzzy framework. The objective of the optimization problem is to minimize the weighted square of the difference between the measured and the estimated quantity. The proposed method of state estimation considers bus voltage magnitude and phase angle as state variable along with some equality and inequality constraints for state estimation in distribution networks. A rule based fuzzy inference system has been designed to control the parameters of the CSS algorithm to achieve better balance between the exploration and exploitation capability of the algorithm. The efficiency of the proposed fuzzy adaptive charged system search (FACSS algorithm has been tested on standard IEEE 33-bus system and Indian 85-bus practical radial distribution system. The obtained results have been compared with the conventional CSS algorithm, weighted least square (WLS algorithm and particle swarm optimization (PSO for feasibility of the algorithm.
On distribution reduction and algorithm implementation in inconsistent ordered information systems.
Zhang, Yanqin
2014-01-01
As one part of our work in ordered information systems, distribution reduction is studied in inconsistent ordered information systems (OISs). Some important properties on distribution reduction are studied and discussed. The dominance matrix is restated for reduction acquisition in dominance relations based information systems. Matrix algorithm for distribution reduction acquisition is stepped. And program is implemented by the algorithm. The approach provides an effective tool for the theoretical research and the applications for ordered information systems in practices. For more detailed and valid illustrations, cases are employed to explain and verify the algorithm and the program which shows the effectiveness of the algorithm in complicated information systems.
Directory of Open Access Journals (Sweden)
Tieyu Zhao
2015-01-01
Full Text Available The optical image encryption has attracted more and more researchers’ attention, and the various encryption schemes have been proposed. In existing optical cryptosystem, the phase functions or images are usually used as the encryption keys, and it is difficult that the traditional public-key algorithm (such as RSA, ECC, etc. is used to complete large numerical key transfer. In this paper, we propose a key distribution scheme based on the phase retrieval algorithm and the RSA public-key algorithm, which solves the problem for the key distribution in optical image encryption system. Furthermore, we also propose a novel image encryption system based on the key distribution principle. In the system, the different keys can be used in every encryption process, which greatly improves the security of the system.
Equalization Algorithm for Distributed Energy Storage Systems in Islanded AC Microgrids
DEFF Research Database (Denmark)
Aldana, Nelson Leonardo Diaz; Hernández, Adriana Carolina Luna; Quintero, Juan Carlos Vasquez
2015-01-01
This paper presents a centralized strategy for equalizing the state of charge of distributed energy storage systems in an islanded ac microgrid. The strategy is based on a simple algorithm denoted as equalization algorithm, which modifies the charge or discharge ratio on the time, for distributed...
Institute of Scientific and Technical Information of China (English)
YANGGuo-Sheng; WENCheng-Lin; TANMin
2004-01-01
A new multisensor distributed track fusion algorithm is put forward based on combiningthe feedback integration with the strong tracking Kalman filter. Firstly, an effective tracking gateis constructed by taking the intersection of the tracking gates formed before and after feedback.Secondly, on the basis of the constructed effective tracking gate, probabilistic data association andstrong tracking Kalman filter are combined to form the new multisensor distributed track fusionalgorithm. At last, simulation is performed on the original algorithm and the algorithm presented.
Agent-based Algorithm for Spatial Distribution of Objects
Collier, Nathan; Sieniek, Marcin
2012-01-01
method. The bubble mesh algorithm treats objects in space as bubbles, which repel and attract each other. The dynamics of each bubble are approximated by solving a series of ordinary differential equations. We present numerical results for a meshing
Algorithms and ordering heuristics for distributed constraint satisfaction problems
Wahbi , Mohamed
2013-01-01
DisCSP (Distributed Constraint Satisfaction Problem) is a general framework for solving distributed problems arising in Distributed Artificial Intelligence.A wide variety of problems in artificial intelligence are solved using the constraint satisfaction problem paradigm. However, there are several applications in multi-agent coordination that are of a distributed nature. In this type of application, the knowledge about the problem, that is, variables and constraints, may be logically or geographically distributed among physical distributed agents. This distribution is mainly due to p
Energy Technology Data Exchange (ETDEWEB)
Niknam, Taher [Electrical and Electronics Engineering Department, Shiraz University of Technology, Modars Blvd. P.O. 71555-313, Shiraz (Iran, Islamic Republic of)
2011-03-15
In recent years, Distributed Generators (DGs) connected to the distribution network have received increasing attention. The connection of enormous DGs into existing distribution network changes the operation of distribution systems. Because of the small X/R ratio and radial structure of distribution systems, DGs affect the daily Volt/Var control. This paper presents a new algorithm for multiobjective daily Volt/Var control in distribution systems including Distributed Generators (DGs). The objectives are costs of energy generation by DGs and distribution companies, electrical energy losses and the voltage deviations for the next day. A new optimization algorithm based on a Chaotic Improved Honey Bee Mating Optimization (CIHBMO) is proposed to determine the active power values of DGs, reactive power values of capacitors and tap positions of transformers for the next day. Since objectives are not the same, a fuzzy system is used to calculate the best solution. The plausibility of the proposed algorithm is demonstrated and its performance is compared with other methods on a 69-bus distribution feeder. Simulation results illustrate that the proposed algorithm has better outperforms the other algorithms. (author)
International Nuclear Information System (INIS)
Niknam, Taher
2011-01-01
In recent years, Distributed Generators (DGs) connected to the distribution network have received increasing attention. The connection of enormous DGs into existing distribution network changes the operation of distribution systems. Because of the small X/R ratio and radial structure of distribution systems, DGs affect the daily Volt/Var control. This paper presents a new algorithm for multiobjective daily Volt/Var control in distribution systems including Distributed Generators (DGs). The objectives are costs of energy generation by DGs and distribution companies, electrical energy losses and the voltage deviations for the next day. A new optimization algorithm based on a Chaotic Improved Honey Bee Mating Optimization (CIHBMO) is proposed to determine the active power values of DGs, reactive power values of capacitors and tap positions of transformers for the next day. Since objectives are not the same, a fuzzy system is used to calculate the best solution. The plausibility of the proposed algorithm is demonstrated and its performance is compared with other methods on a 69-bus distribution feeder. Simulation results illustrate that the proposed algorithm has better outperforms the other algorithms.
A voting-based star identification algorithm utilizing local and global distribution
Fan, Qiaoyun; Zhong, Xuyang; Sun, Junhua
2018-03-01
A novel star identification algorithm based on voting scheme is presented in this paper. In the proposed algorithm, the global distribution and local distribution of sensor stars are fully utilized, and the stratified voting scheme is adopted to obtain the candidates for sensor stars. The database optimization is employed to reduce its memory requirement and improve the robustness of the proposed algorithm. The simulation shows that the proposed algorithm exhibits 99.81% identification rate with 2-pixel standard deviations of positional noises and 0.322-Mv magnitude noises. Compared with two similar algorithms, the proposed algorithm is more robust towards noise, and the average identification time and required memory is less. Furthermore, the real sky test shows that the proposed algorithm performs well on the real star images.
The relationship between randomness and power-law distributed move lengths in random walk algorithms
Sakiyama, Tomoko; Gunji, Yukio-Pegio
2014-05-01
Recently, we proposed a new random walk algorithm, termed the REV algorithm, in which the agent alters the directional rule that governs it using the most recent four random numbers. Here, we examined how a non-bounded number, i.e., "randomness" regarding move direction, was important for optimal searching and power-law distributed step lengths in rule change. We proposed two algorithms: the REV and REV-bounded algorithms. In the REV algorithm, one of the four random numbers used to change the rule is non-bounded. In contrast, all four random numbers in the REV-bounded algorithm are bounded. We showed that the REV algorithm exhibited more consistent power-law distributed step lengths and flexible searching behavior.
Virus evolutionary genetic algorithm for task collaboration of logistics distribution
Ning, Fanghua; Chen, Zichen; Xiong, Li
2005-12-01
In order to achieve JIT (Just-In-Time) level and clients' maximum satisfaction in logistics collaboration, a Virus Evolutionary Genetic Algorithm (VEGA) was put forward under double constraints of logistics resource and operation sequence. Based on mathematic description of a multiple objective function, the algorithm was designed to schedule logistics tasks with different due dates and allocate them to network members. By introducing a penalty item, make span and customers' satisfaction were expressed in fitness function. And a dynamic adaptive probability of infection was used to improve performance of local search. Compared to standard Genetic Algorithm (GA), experimental result illustrates the performance superiority of VEGA. So the VEGA can provide a powerful decision-making technique for optimizing resource configuration in logistics network.
Parallel grid generation algorithm for distributed memory computers
Moitra, Stuti; Moitra, Anutosh
1994-01-01
A parallel grid-generation algorithm and its implementation on the Intel iPSC/860 computer are described. The grid-generation scheme is based on an algebraic formulation of homotopic relations. Methods for utilizing the inherent parallelism of the grid-generation scheme are described, and implementation of multiple levELs of parallelism on multiple instruction multiple data machines are indicated. The algorithm is capable of providing near orthogonality and spacing control at solid boundaries while requiring minimal interprocessor communications. Results obtained on the Intel hypercube for a blended wing-body configuration are used to demonstrate the effectiveness of the algorithm. Fortran implementations bAsed on the native programming model of the iPSC/860 computer and the Express system of software tools are reported. Computational gains in execution time speed-up ratios are given.
Dynamic Consensus Algorithm based Distributed Unbalance Compensation in Islanded Microgrids
DEFF Research Database (Denmark)
Meng, Lexuan; Zhao, Xin; Firoozabadi, Mehdi Savaghebi
2015-01-01
In islanded microgrids (MG), distributed generators (DG) can be employed as distributed compensators for improving the power quality (voltage unbalance and harmonics) in consumer side. Hierarchical control is usually applied with different control levels differentiated. In case of voltage unbalance...
Target distribution in cooperative combat based on Bayesian optimization algorithm
Institute of Scientific and Technical Information of China (English)
Shi Zhifu; Zhang An; Wang Anli
2006-01-01
Target distribution in cooperative combat is a difficult and emphases. We build up the optimization model according to the rule of fire distribution. We have researched on the optimization model with BOA. The BOA can estimate the joint probability distribution of the variables with Bayesian network, and the new candidate solutions also can be generated by the joint distribution. The simulation example verified that the method could be used to solve the complex question, the operation was quickly and the solution was best.
Directory of Open Access Journals (Sweden)
Wanxing Sheng
2014-01-01
Full Text Available To solve the comprehensive multiobjective optimization problem, this study proposes an improved metaheuristic searching algorithm with combination of harmony search and the fast nondominated sorting approach. This is a kind of the novel intelligent optimization algorithm for multiobjective harmony search (MOHS. The detailed description and the algorithm formulating are discussed. Taking the optimal placement and sizing issue of distributed generation (DG in distributed power system as one example, the solving procedure of the proposed method is given. Simulation result on modified IEEE 33-bus test system and comparison with NSGA-II algorithm has proved that the proposed MOHS can get promising results for engineering application.
Distributed interference alignment iterative algorithms in symmetric wireless network
Directory of Open Access Journals (Sweden)
YANG Jingwen
2015-02-01
Full Text Available Interference alignment is a novel interference alignment way,which is widely noted all of the world.Interference alignment overlaps interference in the same signal space at receiving terminal by precoding so as to thoroughly eliminate the influence of interference impacted on expected signals,thus making the desire user achieve the maximum degree of freedom.In this paper we research three typical algorithms for realizing interference alignment,including minimizing the leakage interference,maximizing Signal to Interference plus Noise Ratio (SINR and minimizing mean square error(MSE.All of these algorithms utilize the reciprocity of wireless network,and iterate the precoders between original network and the reverse network so as to achieve interference alignment.We use the uplink transmit rate to analyze the performance of these three algorithms.Numerical simulation results show the advantages of these algorithms.which is the foundation for the further study in the future.The feasibility and future of interference alignment are also discussed at last.
Li, Xiaowei; Mei, Qingqing; Dai, Xiaohu; Ding, Guoji
2017-03-01
Thermogravimetric analysis, Gaussian-fit-peak model (GFPM), and distributed activation energy model (DAEM) were firstly used to explore the effect of anaerobic digestion on sequential pyrolysis kinetic of four organic solid wastes (OSW). Results showed that the OSW weight loss mainly occurred in the second pyrolysis stage relating to organic matter decomposition. Compared with raw substrate, the weight loss of corresponding digestate was lower in the range of 180-550°C, but was higher in 550-900°C. GFPM analysis revealed that organic components volatized at peak temperatures of 188-263, 373-401 and 420-462°C had a faster degradation rate than those at 274-327°C during anaerobic digestion. DAEM analysis showed that anaerobic digestion had discrepant effects on activation energy for four OSW pyrolysis, possibly because of their different organic composition. It requires further investigation for the special organic matter, i.e., protein-like and carbohydrate-like groups, to confirm the assumption. Copyright © 2016 Elsevier Ltd. All rights reserved.
Efficient Algorithms for Distributed Control : A Structured Matrix Approach
Rice, J.K.
2010-01-01
Distributed systems are all around us, and they are fascinating, and have an enormous potential to improve our lives, if their complexity can be properly harnessed. All scientists and engineers are aware of the great potential of this subject, since we witness fantastic distributed control systems
Wei, Yongjie; Ge, Baozhen; Wei, Yaolin
2009-03-20
In general, model-independent algorithms are sensitive to noise during laser particle size measurement. An improved conjugate gradient algorithm (ICGA) that can be used to invert particle size distribution (PSD) from diffraction data is presented. By use of the ICGA to invert simulated data with multiplicative or additive noise, we determined that additive noise is the main factor that induces distorted results. Thus the ICGA is amended by introduction of an iteration step-adjusting parameter and is used experimentally on simulated data and some samples. The experimental results show that the sensitivity of the ICGA to noise is reduced and the inverted results are in accord with the real PSD.
Lower bounds on the run time of the univariate marginal distribution algorithm on OneMax
DEFF Research Database (Denmark)
Krejca, Martin S.; Witt, Carsten
2017-01-01
The Univariate Marginal Distribution Algorithm (UMDA), a popular estimation of distribution algorithm, is studied from a run time perspective. On the classical OneMax benchmark function, a lower bound of Ω(μ√n + n log n), where μ is the population size, on its expected run time is proved...... values maintained by the algorithm, including carefully designed potential functions. These techniques may prove useful in advancing the field of run time analysis for estimation of distribution algorithms in general........ This is the first direct lower bound on the run time of the UMDA. It is stronger than the bounds that follow from general black-box complexity theory and is matched by the run time of many evolutionary algorithms. The results are obtained through advanced analyses of the stochastic change of the frequencies of bit...
Indian Academy of Sciences (India)
polynomial) division have been found in Vedic Mathematics which are dated much before Euclid's algorithm. A programming language Is used to describe an algorithm for execution on a computer. An algorithm expressed using a programming.
Li, Wenhao
2011-01-01
Distributed workflow technology has been widely used in modern education and e-business systems. Distributed web applications have shown cross-domain and cooperative characteristics to meet the need of current distributed workflow applications. In this paper, the author proposes a dynamic and adaptive scheduling algorithm PCSA (Pre-Calculated…
Adaptive Variance Scaling in Continuous Multi-Objective Estimation-of-Distribution Algorithms
P.A.N. Bosman (Peter); D. Thierens (Dirk); D. Thierens (Dirk)
2007-01-01
htmlabstractRecent research into single-objective continuous Estimation-of-Distribution Algorithms (EDAs) has shown that when maximum-likelihood estimations are used for parametric distributions such as the normal distribution, the EDA can easily suffer from premature convergence. In this paper we
Concurrent Learning of Control in Multi agent Sequential Decision Tasks
2018-04-17
Concurrent Learning of Control in Multi-agent Sequential Decision Tasks The overall objective of this project was to develop multi-agent reinforcement... learning (MARL) approaches for intelligent agents to autonomously learn distributed control policies in decentral- ized partially observable... learning of policies in Dec-POMDPs, established performance bounds, evaluated these algorithms both theoretically and empirically, The views
International Nuclear Information System (INIS)
Lerouge, C.; Gaucher, E.C.; Tournassat, C.; Negrel, P.; Crouzet, C.; Guerrot, C.; Gautier, A.; Michel, P.; Vinsot, A.; Buschaert, S.
2010-01-01
Document available in extended abstract form only. Strontium is a minor element in the Callovian-Oxfordian clayey formation of Bure (ANDRA Underground Research laboratory, France) and may be interesting as a natural analogue of cationic radionuclides. The objective of our study was to better understand the chemical behaviour of this element since the sediment deposition. Previous mineralogical work showed that main strontium-bearing minerals in the clay-stones, in addition to clay minerals, are celestite, carbonates and accessory micas and K-feldspars. In order to establish the strontium distribution and determine its origins across the clayey formation and establish a strontium diffusion profile, a four step sequential extraction procedure combined with strontium isotopes was developed and adapted to the mineralogy of the Bure clay-stones. The four-steps sequential extraction developed and applied to sixteen samples of Callovian-Oxfordian clay-stones showed that the exchangeable fraction (27 to 48% of the total strontium) and the carbonate fraction (38-47% of the total strontium) are the two main fractions carrying the strontium, the detrital fraction being secondary (< 15%). Celestite is the major strontium-carrier mineral that can perturb the results of the sequential extraction but not significantly modify the distribution of strontium at the scale of the formation. The celestite was studied on micronic to centi-metric crystals, separated by hand-picking on core samples. The development of this sequential extraction procedure for the Callovian-Oxfordian formation shows the importance of the first cobalt hexamine trichloride step to extract the exchangeable strontium adsorbed on clay minerals and block the exchangeable sites with cobalt for the following steps, allowing a good separation of exchangeable fraction and carbonates. However the reagent/solid ratio of 20 seems to be less adapted to carbonate-rich samples. The good separation between exchangeable
Efficient Reactive Power Compensation Algorithm for Distribution Network
Directory of Open Access Journals (Sweden)
J. Jerome
2017-12-01
Full Text Available The use of automation and energy efficient equipment with electronic control would greatly improve industrial production. These new devices are more sensitive to supply voltage deviation and the characteristics of the power system that was previously ignored are now very important. Hence the benefits of distribution automation have been widely acknowledged in recent years. This paper proposes an efficient load flow solution technique extended to find optimum location for reactive power compensation and network reconfiguration for planning and day-to-day operation of distribution networks. This is required as a part of the distribution automation system (DAS for taking various control and operation decisions. The method exploits the radial nature of the network and uses forward and backward propagation technique to calculate branch currents and node voltages. The proposed method has been tested to analyze several practical distribution networks of various voltage levels and also having high R/X ratio.
Directory of Open Access Journals (Sweden)
Maryam Shokouhi
2017-06-01
Full Text Available Distribution networks are designed as a ring and operated as a radial form. Therefore, the reconfiguration is a simple and cost-effective way to use existing facilities without the need for any new equipment in distribution networks to achieve various objectives such as: power loss reduction, feeder overload reduction, load balancing, voltage profile improvement, reducing the number of switching considering constraints that ultimately result in the power loss reduction. In this paper, a new method based on the Ant Lion algorithm (a modern meta-heuristic algorithm is provided for the reconfiguration of distribution networks. Considering the extension of the distribution networks and complexity of their communications networks, and the various parameters, using smart techniques is inevitable. The proposed approach is tested on the IEEE 33 & 69-bus radial standard distribution networks. The Evaluation of results in MATLAB software shows the effectiveness of the Ant Lion algorithm in the distribution network reconfiguration.
Secure Computation, I/O-Efficient Algorithms and Distributed Signatures
DEFF Research Database (Denmark)
Damgård, Ivan Bjerre; Kölker, Jonas; Toft, Tomas
2012-01-01
values of form r, gr for random secret-shared r ∈ ℤq and gr in a group of order q. This costs a constant number of exponentiation per player per value generated, even if less than n/3 players are malicious. This can be used for efficient distributed computing of Schnorr signatures. We further develop...... the technique so we can sign secret data in a distributed fashion at essentially the same cost....
A Local Scalable Distributed EM Algorithm for Large P2P Networks
National Aeronautics and Space Administration — his paper describes a local and distributed expectation maximization algorithm for learning parameters of Gaussian mixture models (GMM) in large peer-to-peer (P2P)...
A Local Scalable Distributed Expectation Maximization Algorithm for Large Peer-to-Peer Networks
National Aeronautics and Space Administration — This paper describes a local and distributed expectation maximization algorithm for learning parameters of Gaussian mixture models (GMM) in large peer-to-peer (P2P)...
National Aeronautics and Space Administration — In this paper we develop a local distributed privacy preserving algorithm for feature selection in a large peer-to-peer environment. Feature selection is often used...
Directory of Open Access Journals (Sweden)
Kaifeng Yang
2014-01-01
Full Text Available A novel hybrid multiobjective algorithm is presented in this paper, which combines a new multiobjective estimation of distribution algorithm, an efficient local searcher and ε-dominance. Besides, two multiobjective problems with variable linkages strictly based on manifold distribution are proposed. The Pareto set to the continuous multiobjective optimization problems, in the decision space, is a piecewise low-dimensional continuous manifold. The regularity by the manifold features just build probability distribution model by globally statistical information from the population, yet, the efficiency of promising individuals is not well exploited, which is not beneficial to search and optimization process. Hereby, an incremental tournament local searcher is designed to exploit local information efficiently and accelerate convergence to the true Pareto-optimal front. Besides, since ε-dominance is a strategy that can make multiobjective algorithm gain well distributed solutions and has low computational complexity, ε-dominance and the incremental tournament local searcher are combined here. The novel memetic multiobjective estimation of distribution algorithm, MMEDA, was proposed accordingly. The algorithm is validated by experiment on twenty-two test problems with and without variable linkages of diverse complexities. Compared with three state-of-the-art multiobjective optimization algorithms, our algorithm achieves comparable results in terms of convergence and diversity metrics.
Yang, Kaifeng; Mu, Li; Yang, Dongdong; Zou, Feng; Wang, Lei; Jiang, Qiaoyong
2014-01-01
A novel hybrid multiobjective algorithm is presented in this paper, which combines a new multiobjective estimation of distribution algorithm, an efficient local searcher and ε-dominance. Besides, two multiobjective problems with variable linkages strictly based on manifold distribution are proposed. The Pareto set to the continuous multiobjective optimization problems, in the decision space, is a piecewise low-dimensional continuous manifold. The regularity by the manifold features just build probability distribution model by globally statistical information from the population, yet, the efficiency of promising individuals is not well exploited, which is not beneficial to search and optimization process. Hereby, an incremental tournament local searcher is designed to exploit local information efficiently and accelerate convergence to the true Pareto-optimal front. Besides, since ε-dominance is a strategy that can make multiobjective algorithm gain well distributed solutions and has low computational complexity, ε-dominance and the incremental tournament local searcher are combined here. The novel memetic multiobjective estimation of distribution algorithm, MMEDA, was proposed accordingly. The algorithm is validated by experiment on twenty-two test problems with and without variable linkages of diverse complexities. Compared with three state-of-the-art multiobjective optimization algorithms, our algorithm achieves comparable results in terms of convergence and diversity metrics.
EDMC: An enhanced distributed multi-channel anti-collision algorithm for RFID reader system
Zhang, YuJing; Cui, Yinghua
2017-05-01
In this paper, we proposes an enhanced distributed multi-channel reader anti-collision algorithm for RFID environments which is based on the distributed multi-channel reader anti-collision algorithm for RFID environments (called DiMCA). We proposes a monitor method to decide whether reader receive the latest control news after it selected the data channel. The simulation result shows that it improves interrogation delay.
Analysis of Known Linear Distributed Average Consensus Algorithms on Cycles and Paths
Directory of Open Access Journals (Sweden)
Jesús Gutiérrez-Gutiérrez
2018-03-01
Full Text Available In this paper, we compare six known linear distributed average consensus algorithms on a sensor network in terms of convergence time (and therefore, in terms of the number of transmissions required. The selected network topologies for the analysis (comparison are the cycle and the path. Specifically, in the present paper, we compute closed-form expressions for the convergence time of four known deterministic algorithms and closed-form bounds for the convergence time of two known randomized algorithms on cycles and paths. Moreover, we also compute a closed-form expression for the convergence time of the fastest deterministic algorithm considered on grids.
A Nodes Deployment Algorithm in Wireless Sensor Network Based on Distribution
Directory of Open Access Journals (Sweden)
Song Yuli
2014-07-01
Full Text Available Wireless sensor network coverage is a basic problem of wireless sensor network. In this paper, we propose a wireless sensor network node deployment algorithm base on distribution in order to form an efficient wireless sensor network. The iteratively greedy algorithm is used in this paper to choose priority nodes into active until the entire network is covered by wireless sensor nodes, the whole network to multiply connected. The simulation results show that the distributed wireless sensor network node deployment algorithm can form a multiply connected wireless sensor network.
Research on distributed QOS routing algorithm based on TCP/IP
Liu, Xiaoyue; Chen, Yongqiang
2011-10-01
At present, network environment follow protocol standard of IPV4 is intended to do the best effort of network to provide network applied service for users, however, not caring about service quality.Thus the packet loss rate is high, it cannot reach an ideal applied results. This article through the establishment of mathematical model, put forward a new distributed multi QOS routing algorithm, given the realization process of this distributed QOS routing algorithm, and simulation was carried out by simulation software. The results show the proposed algorithm can improve the utilization rate of network resources and the service quality of network application.
Liang, Faming; Jin, Ick-Hoon
2013-01-01
Simulating from distributions with intractable normalizing constants has been a long-standing problem inmachine learning. In this letter, we propose a new algorithm, the Monte Carlo Metropolis-Hastings (MCMH) algorithm, for tackling this problem. The MCMH algorithm is a Monte Carlo version of the Metropolis-Hastings algorithm. It replaces the unknown normalizing constant ratio by a Monte Carlo estimate in simulations, while still converges, as shown in the letter, to the desired target distribution under mild conditions. The MCMH algorithm is illustrated with spatial autologistic models and exponential random graph models. Unlike other auxiliary variable Markov chain Monte Carlo (MCMC) algorithms, such as the Møller and exchange algorithms, the MCMH algorithm avoids the requirement for perfect sampling, and thus can be applied to many statistical models for which perfect sampling is not available or very expensive. TheMCMHalgorithm can also be applied to Bayesian inference for random effect models and missing data problems that involve simulations from a distribution with intractable integrals. © 2013 Massachusetts Institute of Technology.
Liang, Faming
2013-08-01
Simulating from distributions with intractable normalizing constants has been a long-standing problem inmachine learning. In this letter, we propose a new algorithm, the Monte Carlo Metropolis-Hastings (MCMH) algorithm, for tackling this problem. The MCMH algorithm is a Monte Carlo version of the Metropolis-Hastings algorithm. It replaces the unknown normalizing constant ratio by a Monte Carlo estimate in simulations, while still converges, as shown in the letter, to the desired target distribution under mild conditions. The MCMH algorithm is illustrated with spatial autologistic models and exponential random graph models. Unlike other auxiliary variable Markov chain Monte Carlo (MCMC) algorithms, such as the Møller and exchange algorithms, the MCMH algorithm avoids the requirement for perfect sampling, and thus can be applied to many statistical models for which perfect sampling is not available or very expensive. TheMCMHalgorithm can also be applied to Bayesian inference for random effect models and missing data problems that involve simulations from a distribution with intractable integrals. © 2013 Massachusetts Institute of Technology.
Distributed topology control algorithm for multihop wireless netoworks
Borbash, S. A.; Jennings, E. H.
2002-01-01
We present a network initialization algorithmfor wireless networks with distributed intelligence. Each node (agent) has only local, incomplete knowledge and it must make local decisions to meet a predefined global objective. Our objective is to use power control to establish a topology based onthe relative neighborhood graph which has good overall performance in terms of power usage, low interference, and reliability.
A Merging Algorithm for Aerosol Size Distribution from Multiple Instruments
Czech Academy of Sciences Publication Activity Database
Ondráček, Jakub; Ždímal, Vladimír; Smolík, Jiří; Lazaridis, M.
2009-01-01
Roč. 199, 1-4 (2009), s. 219-233 ISSN 0049-6979 Grant - others:MTKD(XE) CT-2004-513849 Institutional research plan: CEZ:AV0Z40720504 Keywords : aerosols * merging particle size distribution * multilognormal model Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 1.676, year: 2009
Agent-aggregator platform and its distributed algorithm
Viyathukattuva Mohamed Ali, M.M.; Nguyen, H.P.; Kling, W.L.
2015-01-01
The activities of the INCREASE project are focused on intelligent ways to manage distributed renewable energy sources (DRES) in low-voltage (LV) and medium-voltage (MV) networks.INCREASE aims at providing solutions to overcome technical issues during power system operation, by employing advanced
Leakage detection algorithm integrating water distribution networks hydraulic model
CSIR Research Space (South Africa)
Adedeji, K
2017-06-01
Full Text Available Water loss through leaking pipes is inexorable in water distribution networks (WDNs) and has been recognized as a major challenge facing the operation of municipal water services. This is strongly linked with financial costs due to economic loss...
A hybrid evolutionary algorithm for distribution feeder reconfiguration
Indian Academy of Sciences (India)
the reconfiguration of distribution networks has been proposed by .... An effective strategy to increase the loading margin of heavily loaded feeders is to ... social animals such as a flock of birds, a school of fish or a group of people that pursue.
Xie, Yanan; Zhou, Mingliang; Pan, Dengke
2017-10-01
The forward-scattering model is introduced to describe the response of normalized radar cross section (NRCS) of precipitation with synthetic aperture radar (SAR). Since the distribution of near-surface rainfall is related to the rate of near-surface rainfall and horizontal distribution factor, a retrieval algorithm called modified regression empirical and model-oriented statistical (M-M) based on the volterra integration theory is proposed. Compared with the model-oriented statistical and volterra integration (MOSVI) algorithm, the biggest difference is that the M-M algorithm is based on the modified regression empirical algorithm rather than the linear regression formula to retrieve the value of near-surface rainfall rate. Half of the empirical parameters are reduced in the weighted integral work and a smaller average relative error is received while the rainfall rate is less than 100 mm/h. Therefore, the algorithm proposed in this paper can obtain high-precision rainfall information.
Generating log-normally distributed random numbers by using the Ziggurat algorithm
International Nuclear Information System (INIS)
Choi, Jong Soo
2016-01-01
Uncertainty analyses are usually based on the Monte Carlo method. Using an efficient random number generator(RNG) is a key element in success of Monte Carlo simulations. Log-normal distributed variates are very typical in NPP PSAs. This paper proposes an approach to generate log normally distributed variates based on the Ziggurat algorithm and evaluates the efficiency of the proposed Ziggurat RNG. The proposed RNG can be helpful to improve the uncertainty analysis of NPP PSAs. This paper focuses on evaluating the efficiency of the Ziggurat algorithm from a NPP PSA point of view. From this study, we can draw the following conclusions. - The Ziggurat algorithm is one of perfect random number generators to product normal distributed variates. - The Ziggurat algorithm is computationally much faster than the most commonly used method, Marsaglia polar method
Bio-Inspired Distributed Decision Algorithms for Anomaly Detection
2017-03-01
Generation Services (ETG) 3. Replay of Traffic Traces (RTT) BTG creates “ norm ” traffic background with pre-specified distribution, BTG takes in a...a cap on the IP counter to offset this artificial effect. For this reason, we also evaluated the dependence of DIAMoND performance on the IP counter... cap . 3.3.2.10 Performance Evaluation Metrics. Given the local anomaly detector is based on TCP session negotiation protocols, it is natural to
Stepwise Development Of Distributed Vertex Coloring Algorithms (Full Report)
Andriamiarina , Manamiary Bruno; Méry , Dominique
2011-01-01
Software-based systems have a strong impact in the daily life. For instance, systems like televisions, cell phones, credit cards are used for persons, while others systems, like networks, telecommunications, distributed and embedded devices, supercomputers, are used by organisations such as companies, governments, nations... Several countries, especially the advanced ones, rely on systems for the efficiency of domains like economy, health... Since they are needed in daily life, those systems ...
Optimization of Nuclear Reactor power Distribution using Genetic Algorithm
International Nuclear Information System (INIS)
Kim, Hyu Chan
1996-02-01
The main purpose of study is to develop a computer code named as 'MGA-SCOUPE' which can determine an optimal fuel-loading pattern for the nuclear reactor. The developed code, MGA-SCOUPE, automatically lots of searches for the globally optimum solutions based upon the modified Genetic Algorithm(MGA). The optimization goal of the MGA-SCOUPE is (1) the minimization of the deviations in the power peaking factors both at BOC and EOC, and (2) the maximization of the average burnup ration at EOC of the total fuel assemblies. For the reactor core calculation module in the MGA-SCOUPE, the SCOUPE code was partially modified and used. It had been developed originally in MIT and has been used currently in Kyung Hee University. The application of the MGA-SCOUPE to KORI 4-4 Cycle Model show several satisfactory results. Among them, two dominant improvements compared with the SCOUPE code can be summarized as follow: - The MGA-SCOUPE removes the user-dependency problem of the SCOUPE in the optimal loading pattern searches. Therefore, the searching process in the MGA-SCOUPE can be easily automated. - The final fuel loading pattern obtained by the MGA-SCOUPE shows 25.8%, 18.7% reduced standard deviations of the power peaking factors both at BOC and EOC, and 45% increased avg. burnup ratio at EOC compare with those of the SCOUPE
Mitavskiy, Boris; Cannings, Chris
2009-01-01
The evolutionary algorithm stochastic process is well-known to be Markovian. These have been under investigation in much of the theoretical evolutionary computing research. When the mutation rate is positive, the Markov chain modeling of an evolutionary algorithm is irreducible and, therefore, has a unique stationary distribution. Rather little is known about the stationary distribution. In fact, the only quantitative facts established so far tell us that the stationary distributions of Markov chains modeling evolutionary algorithms concentrate on uniform populations (i.e., those populations consisting of a repeated copy of the same individual). At the same time, knowing the stationary distribution may provide some information about the expected time it takes for the algorithm to reach a certain solution, assessment of the biases due to recombination and selection, and is of importance in population genetics to assess what is called a "genetic load" (see the introduction for more details). In the recent joint works of the first author, some bounds have been established on the rates at which the stationary distribution concentrates on the uniform populations. The primary tool used in these papers is the "quotient construction" method. It turns out that the quotient construction method can be exploited to derive much more informative bounds on ratios of the stationary distribution values of various subsets of the state space. In fact, some of the bounds obtained in the current work are expressed in terms of the parameters involved in all the three main stages of an evolutionary algorithm: namely, selection, recombination, and mutation.
Experiments with the auction algorithm for the shortest path problem
DEFF Research Database (Denmark)
Larsen, Jesper; Pedersen, Ib
1999-01-01
The auction approach for the shortest path problem (SPP) as introduced by Bertsekas is tested experimentally. Parallel algorithms using the auction approach are developed and tested. Both the sequential and parallel auction algorithms perform significantly worse than a state-of-the-art Dijkstra-l......-like reference algorithm. Experiments are run on a distributed-memory MIMD class Meiko parallel computer....
Speedup of minimum discontinuity phase unwrapping algorithm with a reference phase distribution
Liu, Yihang; Han, Yu; Li, Fengjiao; Zhang, Qican
2018-06-01
In three-dimensional (3D) shape measurement based on phase analysis, the phase analysis process usually produces a wrapped phase map ranging from - π to π with some 2 π discontinuities, and thus a phase unwrapping algorithm is necessary to recover the continuous and nature phase map from which 3D height distribution can be restored. Usually, the minimum discontinuity phase unwrapping algorithm can be used to solve many different kinds of phase unwrapping problems, but its main drawback is that it requires a large amount of computations and has low efficiency in searching for the improving loop within the phase's discontinuity area. To overcome this drawback, an improvement to speedup of the minimum discontinuity phase unwrapping algorithm by using the phase distribution on reference plane is proposed. In this improved algorithm, before the minimum discontinuity phase unwrapping algorithm is carried out to unwrap phase, an integer number K was calculated from the ratio of the wrapped phase to the nature phase on a reference plane. And then the jump counts of the unwrapped phase can be reduced by adding 2K π, so the efficiency of the minimum discontinuity phase unwrapping algorithm is significantly improved. Both simulated and experimental data results verify the feasibility of the proposed improved algorithm, and both of them clearly show that the algorithm works very well and has high efficiency.
International Nuclear Information System (INIS)
Azadeh, A.; Ghaderi, S.F.; Omrani, H.; Eivazy, H.
2009-01-01
This paper presents an integrated data envelopment analysis (DEA)-corrected ordinary least squares (COLS)-stochastic frontier analysis (SFA)-principal component analysis (PCA)-numerical taxonomy (NT) algorithm for performance assessment, optimization and policy making of electricity distribution units. Previous studies have generally used input-output DEA models for benchmarking and evaluation of electricity distribution units. However, this study proposes an integrated flexible approach to measure the rank and choose the best version of the DEA method for optimization and policy making purposes. It covers both static and dynamic aspects of information environment due to involvement of SFA which is finally compared with the best DEA model through the Spearman correlation technique. The integrated approach would yield in improved ranking and optimization of electricity distribution systems. To illustrate the usability and reliability of the proposed algorithm, 38 electricity distribution units in Iran have been considered, ranked and optimized by the proposed algorithm of this study.
Indian Academy of Sciences (India)
to as 'divide-and-conquer'. Although there has been a large effort in realizing efficient algorithms, there are not many universally accepted algorithm design paradigms. In this article, we illustrate algorithm design techniques such as balancing, greedy strategy, dynamic programming strategy, and backtracking or traversal of ...
Directory of Open Access Journals (Sweden)
Marco Laubacher
2016-06-01
Full Text Available Spatially distributed sequential stimulation (SDSS has demonstrated substantial power output and fatigue benefits compared to single electrode stimulation (SES in the application of functional electrical stimulation (FES. This asymmetric electrode setup brings new possibilities but also new questions since precise placement of the electrodes is one critical factor for good muscle activation. The aim of this study was to compare the power output, fatigue and activation properties of proximally versus distally placed SDSS electrodes in an isokinetic knee extension task simulating knee movement during recumbent cycling. M. vastus lateralis and medialis of seven able-bodied subjects were stimulated with rectangular bi-phasic pulses of constant amplitude of 40 mA and at an SDSS frequency of 35 Hz for 6 min on both legs with both setups (i.e. n=14. Torque was measured during knee-extension movement by a dynamometer at an angular velocity of 110 deg/s. Mean power, peak power and activation time were calculated and compared for the initial and final stimulation phases, together with an overall fatigue index. Power output values (Pmean, Ppeak were scaled to a standardised reference input pulse width of 100 μs (Pmean,s, Ppeak,s. The initial evaluation phase showed no significant differences between the two setups for all outcome measures. Ppeak and Ppeak,s were both significantly higher in the final phase for the distal setup (25.4 ± 8.1 W vs. 28.2 ± 6.2 W, p=0.0062 and 34.8 ± 9.5 W vs. 38.9 ± 6.7 W, p=0.021, respectively. With distal SDSS, there was modest evidence of higher Pmean and Pmean,s (p=0.071, p=0.14, respectively but of longer activation time (p=0.096. The rate of fatigue was similar for both setups. For practical FES applications, distal placement of the SDSS electrodes is preferable.
International Nuclear Information System (INIS)
Cierjacks, S.W.; Oblozinsky, P.; Kelzenberg, S.; Rzehorz, B.
1993-01-01
A new algorithm and three major nuclear data libraries were developed for the kinematically complete treatment of sequential (x,n) reactions in fusion material activation calculations. The new libraries include data for virtually all isotopes with Z ≤ 84 (A ≤ 210) and half-lives exceeding 1 day; primary neutron energies E n 3 He, and α with energies E x < 24 MeV. While production cross sections of charged particles for primary (n,x) reactions can be deduced from the European activation file, the KFKSPEC data file was created for the corresponding normalized charged-particle spectra. The second data file, KFKXN, contains cross sections for secondary (x,n) reactions. The third data file, KFKSTOP, has a complete set of differential ranges for all five aforementioned light charged particles and all elements from hydrogen to uranium. The KFKSPEC and KFKXN libraries are based essentially on nuclear model calculations using the statistical evaporation model superimposed with the pre-equilibrium contribution as implemented in the Lawrence Livermore National Laboratory ALICE code. The KFKSPEC library includes 633 isotopes, of which 55 are in their isomeric states, and contains 63,300 spectra of the (n,x) type with almost 1.5 million data points. The KFKXN library also includes 633 isotopes and contains all (x,n) and partly (x,2n) cross sections for 4431 reactions with ∼ 106,000 data points. The KFKSTOP library is considered complete and has 11,040 data points. 42 refs., 2 figs., 4 tabs
Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.
Directory of Open Access Journals (Sweden)
Shaoming Pan
Full Text Available Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.
Distributed Storage Algorithm for Geospatial Image Data Based on Data Access Patterns.
Pan, Shaoming; Li, Yongkai; Xu, Zhengquan; Chong, Yanwen
2015-01-01
Declustering techniques are widely used in distributed environments to reduce query response time through parallel I/O by splitting large files into several small blocks and then distributing those blocks among multiple storage nodes. Unfortunately, however, many small geospatial image data files cannot be further split for distributed storage. In this paper, we propose a complete theoretical system for the distributed storage of small geospatial image data files based on mining the access patterns of geospatial image data using their historical access log information. First, an algorithm is developed to construct an access correlation matrix based on the analysis of the log information, which reveals the patterns of access to the geospatial image data. Then, a practical heuristic algorithm is developed to determine a reasonable solution based on the access correlation matrix. Finally, a number of comparative experiments are presented, demonstrating that our algorithm displays a higher total parallel access probability than those of other algorithms by approximately 10-15% and that the performance can be further improved by more than 20% by simultaneously applying a copy storage strategy. These experiments show that the algorithm can be applied in distributed environments to help realize parallel I/O and thereby improve system performance.
Automatic Regionalization Algorithm for Distributed State Estimation in Power Systems: Preprint
Energy Technology Data Exchange (ETDEWEB)
Wang, Dexin; Yang, Liuqing; Florita, Anthony; Alam, S.M. Shafiul; Elgindy, Tarek; Hodge, Bri-Mathias
2016-08-01
The deregulation of the power system and the incorporation of generation from renewable energy sources recessitates faster state estimation in the smart grid. Distributed state estimation (DSE) has become a promising and scalable solution to this urgent demand. In this paper, we investigate the regionalization algorithms for the power system, a necessary step before distributed state estimation can be performed. To the best of the authors' knowledge, this is the first investigation on automatic regionalization (AR). We propose three spectral clustering based AR algorithms. Simulations show that our proposed algorithms outperform the two investigated manual regionalization cases. With the help of AR algorithms, we also show how the number of regions impacts the accuracy and convergence speed of the DSE and conclude that the number of regions needs to be chosen carefully to improve the convergence speed of DSEs.
Directory of Open Access Journals (Sweden)
J. Trdlicka
2010-12-01
Full Text Available This work proposes a distributed algorithm for energy optimal routing in a wireless sensor network. The routing problem is described as a mathematical problem by the minimum-cost multi-commodity network flow problem. Due to the separability of the problem, we use the duality theorem to derive the distributed algorithm. The algorithm computes the energy optimal routing in the network without any central node or knowledge of the whole network structure. Each node only needs to know the flow which is supposed to send or receive and the costs and capacities of the neighboring links. An evaluation of the presented algorithm on benchmarks for the energy optimal data flow routing in sensor networks with up to 100 nodes is presented.
Exploring design tradeoffs of a distributed algorithm for cosmic ray event detection
Yousaf, S.; Bakhshi, R.; van Steen, M.; Voulgaris, S.; Kelley, J. L.
2013-03-01
Many sensor networks, including large particle detector arrays measuring high-energy cosmic-ray air showers, traditionally rely on centralised trigger algorithms to find spatial and temporal coincidences of individual nodes. Such schemes suffer from scalability problems, especially if the nodes communicate wirelessly or have bandwidth limitations. However, nodes which instead communicate with each other can, in principle, use a distributed algorithm to find coincident events themselves without communication with a central node. We present such an algorithm and consider various design tradeoffs involved, in the context of a potential trigger for the Auger Engineering Radio Array (AERA).
A Fast Elitism Gaussian Estimation of Distribution Algorithm and Application for PID Optimization
Directory of Open Access Journals (Sweden)
Qingyang Xu
2014-01-01
Full Text Available Estimation of distribution algorithm (EDA is an intelligent optimization algorithm based on the probability statistics theory. A fast elitism Gaussian estimation of distribution algorithm (FEGEDA is proposed in this paper. The Gaussian probability model is used to model the solution distribution. The parameters of Gaussian come from the statistical information of the best individuals by fast learning rule. A fast learning rule is used to enhance the efficiency of the algorithm, and an elitism strategy is used to maintain the convergent performance. The performances of the algorithm are examined based upon several benchmarks. In the simulations, a one-dimensional benchmark is used to visualize the optimization process and probability model learning process during the evolution, and several two-dimensional and higher dimensional benchmarks are used to testify the performance of FEGEDA. The experimental results indicate the capability of FEGEDA, especially in the higher dimensional problems, and the FEGEDA exhibits a better performance than some other algorithms and EDAs. Finally, FEGEDA is used in PID controller optimization of PMSM and compared with the classical-PID and GA.
A fast elitism Gaussian estimation of distribution algorithm and application for PID optimization.
Xu, Qingyang; Zhang, Chengjin; Zhang, Li
2014-01-01
Estimation of distribution algorithm (EDA) is an intelligent optimization algorithm based on the probability statistics theory. A fast elitism Gaussian estimation of distribution algorithm (FEGEDA) is proposed in this paper. The Gaussian probability model is used to model the solution distribution. The parameters of Gaussian come from the statistical information of the best individuals by fast learning rule. A fast learning rule is used to enhance the efficiency of the algorithm, and an elitism strategy is used to maintain the convergent performance. The performances of the algorithm are examined based upon several benchmarks. In the simulations, a one-dimensional benchmark is used to visualize the optimization process and probability model learning process during the evolution, and several two-dimensional and higher dimensional benchmarks are used to testify the performance of FEGEDA. The experimental results indicate the capability of FEGEDA, especially in the higher dimensional problems, and the FEGEDA exhibits a better performance than some other algorithms and EDAs. Finally, FEGEDA is used in PID controller optimization of PMSM and compared with the classical-PID and GA.
Packets Distributing Evolutionary Algorithm Based on PSO for Ad Hoc Network
Xu, Xiao-Feng
2018-03-01
Wireless communication network has such features as limited bandwidth, changeful channel and dynamic topology, etc. Ad hoc network has lots of difficulties in accessing control, bandwidth distribution, resource assign and congestion control. Therefore, a wireless packets distributing Evolutionary algorithm based on PSO (DPSO)for Ad Hoc Network is proposed. Firstly, parameters impact on performance of network are analyzed and researched to obtain network performance effective function. Secondly, the improved PSO Evolutionary Algorithm is used to solve the optimization problem from local to global in the process of network packets distributing. The simulation results show that the algorithm can ensure fairness and timeliness of network transmission, as well as improve ad hoc network resource integrated utilization efficiency.
Directory of Open Access Journals (Sweden)
J. L. Guardado
2014-01-01
Full Text Available Network reconfiguration is an alternative to reduce power losses and optimize the operation of power distribution systems. In this paper, an encoding scheme for evolutionary algorithms is proposed in order to search efficiently for the Pareto-optimal solutions during the reconfiguration of power distribution systems considering multiobjective optimization. The encoding scheme is based on the edge window decoder (EWD technique, which was embedded in the Strength Pareto Evolutionary Algorithm 2 (SPEA2 and the Nondominated Sorting Genetic Algorithm II (NSGA-II. The effectiveness of the encoding scheme was proved by solving a test problem for which the true Pareto-optimal solutions are known in advance. In order to prove the practicability of the encoding scheme, a real distribution system was used to find the near Pareto-optimal solutions for different objective functions to optimize.
Distributed consensus for metamorphic systems using a gossip algorithm for CAT(0) metric spaces
Bellachehab, Anass; Jakubowicz, Jérémie
2015-01-01
We present an application of distributed consensus algorithms to metamorphic systems. A metamorphic system is a set of identical units that can self-assemble to form a rigid structure. For instance, one can think of a robotic arm composed of multiple links connected by joints. The system can change its shape in order to adapt to different environments via reconfiguration of its constituting units. We assume in this work that several metamorphic systems form a network: two systems are connected whenever they are able to communicate with each other. The aim of this paper is to propose a distributed algorithm that synchronizes all the systems in the network. Synchronizing means that all the systems should end up having the same configuration. This aim is achieved in two steps: (i) we cast the problem as a consensus problem on a metric space and (ii) we use a recent distributed consensus algorithm that only make use of metrical notions.
Guardado, J L; Rivas-Davalos, F; Torres, J; Maximov, S; Melgoza, E
2014-01-01
Network reconfiguration is an alternative to reduce power losses and optimize the operation of power distribution systems. In this paper, an encoding scheme for evolutionary algorithms is proposed in order to search efficiently for the Pareto-optimal solutions during the reconfiguration of power distribution systems considering multiobjective optimization. The encoding scheme is based on the edge window decoder (EWD) technique, which was embedded in the Strength Pareto Evolutionary Algorithm 2 (SPEA2) and the Nondominated Sorting Genetic Algorithm II (NSGA-II). The effectiveness of the encoding scheme was proved by solving a test problem for which the true Pareto-optimal solutions are known in advance. In order to prove the practicability of the encoding scheme, a real distribution system was used to find the near Pareto-optimal solutions for different objective functions to optimize.
Shao, Zhongshi; Pi, Dechang; Shao, Weishi
2017-11-01
This article proposes an extended continuous estimation of distribution algorithm (ECEDA) to solve the permutation flow-shop scheduling problem (PFSP). In ECEDA, to make a continuous estimation of distribution algorithm (EDA) suitable for the PFSP, the largest order value rule is applied to convert continuous vectors to discrete job permutations. A probabilistic model based on a mixed Gaussian and Cauchy distribution is built to maintain the exploration ability of the EDA. Two effective local search methods, i.e. revolver-based variable neighbourhood search and Hénon chaotic-based local search, are designed and incorporated into the EDA to enhance the local exploitation. The parameters of the proposed ECEDA are calibrated by means of a design of experiments approach. Simulation results and comparisons based on some benchmark instances show the efficiency of the proposed algorithm for solving the PFSP.
A new stochastic algorithm for inversion of dust aerosol size distribution
Wang, Li; Li, Feng; Yang, Ma-ying
2015-08-01
Dust aerosol size distribution is an important source of information about atmospheric aerosols, and it can be determined from multiwavelength extinction measurements. This paper describes a stochastic inverse technique based on artificial bee colony (ABC) algorithm to invert the dust aerosol size distribution by light extinction method. The direct problems for the size distribution of water drop and dust particle, which are the main elements of atmospheric aerosols, are solved by the Mie theory and the Lambert-Beer Law in multispectral region. And then, the parameters of three widely used functions, i.e. the log normal distribution (L-N), the Junge distribution (J-J), and the normal distribution (N-N), which can provide the most useful representation of aerosol size distributions, are inversed by the ABC algorithm in the dependent model. Numerical results show that the ABC algorithm can be successfully applied to recover the aerosol size distribution with high feasibility and reliability even in the presence of random noise.
DEFF Research Database (Denmark)
Meng, Lexuan; Dragicevic, Tomislav; Roldan Perez, Javier
2016-01-01
Distributed control methods based on consensus algorithms have become popular in recent years for microgrid (MG) systems. These kinds of algorithms can be applied to share information in order to coordinate multiple distributed generators within a MG. However, stability analysis becomes a challen......Distributed control methods based on consensus algorithms have become popular in recent years for microgrid (MG) systems. These kinds of algorithms can be applied to share information in order to coordinate multiple distributed generators within a MG. However, stability analysis becomes...... in the communication network, continuous-time methods can be inaccurate for this kind of dynamic study. Therefore, this paper aims at modeling a complete DC MG using a discrete-time approach in order to perform a sensitivity analysis taking into account the effects of the consensus algorithm. To this end......, a generalized modeling method is proposed and the influence of key control parameters, the communication topology and the communication speed are studied in detail. The theoretical results obtained with the proposed model are verified by comparing them with the results obtained with a detailed switching...
Optimal siting of capacitors in radial distribution network using Whale Optimization Algorithm
Directory of Open Access Journals (Sweden)
D.B. Prakash
2017-12-01
Full Text Available In present days, continuous effort is being made in bringing down the line losses of the electrical distribution networks. Therefore proper allocation of capacitors is of utmost importance because, it will help in reducing the line losses and maintaining the bus voltage. This in turn results in improving the stability and reliability of the system. In this paper Whale Optimization Algorithm (WOA is used to find optimal sizing and placement of capacitors for a typical radial distribution system. Multi objectives such as operating cost reduction and power loss minimization with inequality constraints on voltage limits are considered and the proposed algorithm is validated by applying it on standard radial systems: IEEE-34 bus and IEEE-85 bus radial distribution test systems. The results obtained are compared with those of existing algorithms. The results show that the proposed algorithm is more effective in bringing down the operating costs and in maintaining better voltage profile. Keywords: Whale Optimization Algorithm (WOA, Optimal allocation and sizing of capacitors, Power loss reduction and voltage stability improvement, Radial distribution system, Operating cost minimization
Research on distributed heterogeneous data PCA algorithm based on cloud platform
Zhang, Jin; Huang, Gang
2018-05-01
Principal component analysis (PCA) of heterogeneous data sets can solve the problem that centralized data scalability is limited. In order to reduce the generation of intermediate data and error components of distributed heterogeneous data sets, a principal component analysis algorithm based on heterogeneous data sets under cloud platform is proposed. The algorithm performs eigenvalue processing by using Householder tridiagonalization and QR factorization to calculate the error component of the heterogeneous database associated with the public key to obtain the intermediate data set and the lost information. Experiments on distributed DBM heterogeneous datasets show that the model method has the feasibility and reliability in terms of execution time and accuracy.
An efficient algorithm for generating random number pairs drawn from a bivariate normal distribution
Campbell, C. W.
1983-01-01
An efficient algorithm for generating random number pairs from a bivariate normal distribution was developed. Any desired value of the two means, two standard deviations, and correlation coefficient can be selected. Theoretically the technique is exact and in practice its accuracy is limited only by the quality of the uniform distribution random number generator, inaccuracies in computer function evaluation, and arithmetic. A FORTRAN routine was written to check the algorithm and good accuracy was obtained. Some small errors in the correlation coefficient were observed to vary in a surprisingly regular manner. A simple model was developed which explained the qualities aspects of the errors.
A simple algorithm for measuring particle size distributions on an uneven background from TEM images
DEFF Research Database (Denmark)
Gontard, Lionel Cervera; Ozkaya, Dogan; Dunin-Borkowski, Rafal E.
2011-01-01
Nanoparticles have a wide range of applications in science and technology. Their sizes are often measured using transmission electron microscopy (TEM) or X-ray diffraction. Here, we describe a simple computer algorithm for measuring particle size distributions from TEM images in the presence of a...... application to images of heterogeneous catalysts is presented.......Nanoparticles have a wide range of applications in science and technology. Their sizes are often measured using transmission electron microscopy (TEM) or X-ray diffraction. Here, we describe a simple computer algorithm for measuring particle size distributions from TEM images in the presence...
Energy Technology Data Exchange (ETDEWEB)
Niknam, Taher [Electronic and Electrical Engineering Department, Shiraz University of Technology, Shiraz (Iran)
2009-08-15
This paper introduces a robust searching hybrid evolutionary algorithm to solve the multi-objective Distribution Feeder Reconfiguration (DFR). The main objective of the DFR is to minimize the real power loss, deviation of the nodes' voltage, the number of switching operations, and balance the loads on the feeders. Because of the fact that the objectives are different and no commensurable, it is difficult to solve the problem by conventional approaches that may optimize a single objective. This paper presents a new approach based on norm3 for the DFR problem. In the proposed method, the objective functions are considered as a vector and the aim is to maximize the distance (norm2) between the objective function vector and the worst objective function vector while the constraints are met. Since the proposed DFR is a multi objective and non-differentiable optimization problem, a new hybrid evolutionary algorithm (EA) based on the combination of the Honey Bee Mating Optimization (HBMO) and the Discrete Particle Swarm Optimization (DPSO), called DPSO-HBMO, is implied to solve it. The results of the proposed reconfiguration method are compared with the solutions obtained by other approaches, the original DPSO and HBMO over different distribution test systems. (author)
Cherkashin, N.; Daghbouj, N.; Seine, G.; Claverie, A.
2018-04-01
Sequential He++H+ ion implantation, being more effective than the sole implantation of H+ or He+, is used by many to transfer thin layers of silicon onto different substrates. However, due to the poor understanding of the basic mechanisms involved in such a process, the implantation parameters to be used for the efficient delamination of a superficial layer are still subject to debate. In this work, by using various experimental techniques, we have studied the influence of the He and H relative depth-distributions imposed by the ion energies onto the result of the sequential implantation and annealing of the same fluence of He and H ions. Analyzing the characteristics of the blister populations observed after annealing and deducing the composition of the gas they contain from FEM simulations, we show that the trapping efficiency of He atoms in platelets and blisters during annealing depends on the behavior of the vacancies generated by the two implants within the H-rich region before and after annealing. Maximum efficiency of the sequential ion implantation is obtained when the H-rich region is able to trap all implanted He ions, while the vacancies it generated are not available to favor the formation of V-rich complexes after implantation then He-filled nano-bubbles after annealing. A technological option is to implant He+ ions first at such an energy that the damage it generates is located on the deeper side of the H profile.
An efficient central DOA tracking algorithm for multiple incoherently distributed sources
Hassen, Sonia Ben; Samet, Abdelaziz
2015-12-01
In this paper, we develop a new tracking method for the direction of arrival (DOA) parameters assuming multiple incoherently distributed (ID) sources. The new approach is based on a simple covariance fitting optimization technique exploiting the central and noncentral moments of the source angular power densities to estimate the central DOAs. The current estimates are treated as measurements provided to the Kalman filter that model the dynamic property of directional changes for the moving sources. Then, the covariance-fitting-based algorithm and the Kalman filtering theory are combined to formulate an adaptive tracking algorithm. Our algorithm is compared to the fast approximated power iteration-total least square-estimation of signal parameters via rotational invariance technique (FAPI-TLS-ESPRIT) algorithm using the TLS-ESPRIT method and the subspace updating via FAPI-algorithm. It will be shown that the proposed algorithm offers an excellent DOA tracking performance and outperforms the FAPI-TLS-ESPRIT method especially at low signal-to-noise ratio (SNR) values. Moreover, the performances of the two methods increase as the SNR values increase. This increase is more prominent with the FAPI-TLS-ESPRIT method. However, their performances degrade when the number of sources increases. It will be also proved that our method depends on the form of the angular distribution function when tracking the central DOAs. Finally, it will be shown that the more the sources are spaced, the more the proposed method can exactly track the DOAs.
Study on distributed re-clustering algorithm for moblie wireless sensor networks
Directory of Open Access Journals (Sweden)
XU Chaojie
2016-04-01
Full Text Available In mobile wireless sensor networks,node mobility influences the topology of the hierarchically clustered network,thus affects packet delivery ratio and energy consumption of communications in clusters.To reduce the influence of node mobility,a distributed re-clustering algorithm is proposed in this paper.In this algorithm,basing on the clustered network,nodes estimate their current locations with particle algorithm and predict the most possible locations of next time basing on the mobility model.Each boundary node of a cluster periodically estimates the need for re-clustering and re-cluster itself to the optimal cluster through communicating with the cluster headers when needed.The simulation results indicate that,with small re-clustering periods,the proposed algorithm can be effective to keep appropriate communication distance and outperforms existing schemes on packet delivery ratio and energy consumption.
Research and application of genetic algorithm in path planning of logistics distribution vehicle
Wang, Yong; Zhou, Heng; Wang, Ying
2017-08-01
The core of the logistics distribution system is the vehicle routing planning, research path planning problem, provide a better solution has become an important issue. In order to provide the decision support for logistics and distribution operations, this paper studies the problem of vehicle routing with capacity constraints (CVRP). By establishing a mathematical model, the genetic algorithm is used to plan the path of the logistics vehicle to meet the minimum logistics and transportation costs.
DEFF Research Database (Denmark)
Awasthi, Abhishek; Venkitusamy, Karthikeyan; Padmanaban, Sanjeevikumar
2017-01-01
India's ever increasing population has made it necessary to develop alternative modes of transportation with electric vehicles being the most preferred option. The major obstacle is the deteriorating impact on the utility distribution system brought about by improper setup of these charging...... stations. This paper deals with the optimal planning (siting and sizing) of charging station infrastructure in the city of Allahabad, India. This city is one of the upcoming smart cities, where electric vehicle transportation pilot project is going on under Government of India initiative. In this context......, a hybrid algorithm based on genetic algorithm and improved version of conventional particle swarm optimization is utilized for finding optimal placement of charging station in the Allahabad distribution system. The particle swarm optimization algorithm re-optimizes the received sub-optimal solution (site...
Study of Power Flow Algorithm of AC/DC Distribution System including VSC-MTDC
Directory of Open Access Journals (Sweden)
Haifeng Liang
2015-08-01
Full Text Available In recent years, distributed generation and a large number of sensitive AC and DC loads have been connected to distribution networks, which introduce a series of challenges to distribution network operators (DNOs. In addition, the advantages of DC distribution networks, such as the energy conservation and emission reduction, mean that the voltage source converter based multi-terminal direct current (VSC-MTDC for AC/DC distribution systems demonstrates a great potential, hence drawing growing research interest. In this paper, considering losses of the reactor, the filter and the converter, a mathematical model of VSC-HVDC for the load flow analysis is derived. An AC/DC distribution network architecture has been built, based on which the differences in modified equations of the VSC-MTDC-based network under different control modes are analyzed. In addition, corresponding interface functions under five control modes are provided, and a back/forward iterative algorithm which is applied to power flow calculation of the AC/DC distribution system including VSC-MTDC is proposed. Finally, by calculating the power flow of the modified IEEE14 AC/DC distribution network, the efficiency and validity of the model and algorithm are evaluated. With various distributed generations connected to the network at appropriate locations, power flow results show that network losses and utilization of transmission networks are effectively reduced.
Multi-objective optimization with estimation of distribution algorithm in a noisy environment.
Shim, Vui Ann; Tan, Kay Chen; Chia, Jun Yong; Al Mamun, Abdullah
2013-01-01
Many real-world optimization problems are subjected to uncertainties that may be characterized by the presence of noise in the objective functions. The estimation of distribution algorithm (EDA), which models the global distribution of the population for searching tasks, is one of the evolutionary computation techniques that deals with noisy information. This paper studies the potential of EDAs; particularly an EDA based on restricted Boltzmann machines that handles multi-objective optimization problems in a noisy environment. Noise is introduced to the objective functions in the form of a Gaussian distribution. In order to reduce the detrimental effect of noise, a likelihood correction feature is proposed to tune the marginal probability distribution of each decision variable. The EDA is subsequently hybridized with a particle swarm optimization algorithm in a discrete domain to improve its search ability. The effectiveness of the proposed algorithm is examined via eight benchmark instances with different characteristics and shapes of the Pareto optimal front. The scalability, hybridization, and computational time are rigorously studied. Comparative studies show that the proposed approach outperforms other state of the art algorithms.
COM-LOC: A Distributed Range-Free Localization Algorithm in Wireless Networks
Dil, B.J.; Havinga, Paul J.M.; Marusic, S; Palaniswami, M; Gubbi, J.; Law, Y.W.
2009-01-01
This paper investigates distributed range-free localization in wireless networks using a communication protocol called sum-dist which is commonly employed by localization algorithms. With this protocol, the reference nodes flood the network in order to estimate the shortest distance between the
Indian Academy of Sciences (India)
ticians but also forms the foundation of computer science. Two ... with methods of developing algorithms for solving a variety of problems but ... applications of computers in science and engineer- ... numerical calculus are as important. We will ...
Zhang, Yu; Li, Fei; Zhang, Shengkai; Zhu, Tingting
2017-04-01
Synthetic Aperture Radar (SAR) is significantly important for polar remote sensing since it can provide continuous observations in all days and all weather. SAR can be used for extracting the surface roughness information characterized by the variance of dielectric properties and different polarization channels, which make it possible to observe different ice types and surface structure for deformation analysis. In November, 2016, Chinese National Antarctic Research Expedition (CHINARE) 33rd cruise has set sails in sea ice zone in Antarctic. Accurate leads spatial distribution in sea ice zone for routine planning of ship navigation is essential. In this study, the semantic relationship between leads and sea ice categories has been described by the Conditional Random Fields (CRF) model, and leads characteristics have been modeled by statistical distributions in SAR imagery. In the proposed algorithm, a mixture statistical distribution based CRF is developed by considering the contexture information and the statistical characteristics of sea ice for improving leads detection in Sentinel-1A dual polarization SAR imagery. The unary potential and pairwise potential in CRF model is constructed by integrating the posteriori probability estimated from statistical distributions. For mixture statistical distribution parameter estimation, Method of Logarithmic Cumulants (MoLC) is exploited for single statistical distribution parameters estimation. The iteration based Expectation Maximal (EM) algorithm is investigated to calculate the parameters in mixture statistical distribution based CRF model. In the posteriori probability inference, graph-cut energy minimization method is adopted in the initial leads detection. The post-processing procedures including aspect ratio constrain and spatial smoothing approaches are utilized to improve the visual result. The proposed method is validated on Sentinel-1A SAR C-band Extra Wide Swath (EW) Ground Range Detected (GRD) imagery with a
Distributed routing algorithms to manage power flow in agent-based active distribution network
Nguyen, H.P.; Kling, W.L.; Georgiadis, G.; Papatriantafilou, M.; Anh-Tuan, L.; Bertling, L.
2010-01-01
The current transition from passive to active electric distribution networks comes with problems and challenges on bi-directional power flow in the network and the uncertainty in the forecast of power generation from grid-connected renewable and distributed energy sources. The power flow management
A solution for automatic parallelization of sequential assembly code
Directory of Open Access Journals (Sweden)
Kovačević Đorđe
2013-01-01
Full Text Available Since modern multicore processors can execute existing sequential programs only on a single core, there is a strong need for automatic parallelization of program code. Relying on existing algorithms, this paper describes one new software solution tool for parallelization of sequential assembly code. The main goal of this paper is to develop the parallelizator which reads sequential assembler code and at the output provides parallelized code for MIPS processor with multiple cores. The idea is the following: the parser translates assembler input file to program objects suitable for further processing. After that the static single assignment is done. Based on the data flow graph, the parallelization algorithm separates instructions on different cores. Once sequential code is parallelized by the parallelization algorithm, registers are allocated with the algorithm for linear allocation, and the result at the end of the program is distributed assembler code on each of the cores. In the paper we evaluate the speedup of the matrix multiplication example, which was processed by the parallelizator of assembly code. The result is almost linear speedup of code execution, which increases with the number of cores. The speed up on the two cores is 1.99, while on 16 cores the speed up is 13.88.
International Nuclear Information System (INIS)
Thakur, Amit; Singh, Baltej; Gupta, Anurag; Duggal, Vibhuti; Bhatt, Kislay; Krishnani, P.D.
2016-01-01
Highlights: • EDA has been applied to optimize initial core of AHWR-LEU. • Suitable value of weighing factor ‘α’ and population size in EDA was estimated. • The effect of varying initial distribution function on optimized solution was studied. • For comparison, Genetic algorithm was also applied. - Abstract: Population based evolutionary algorithms now form an integral part of fuel management in nuclear reactors and are frequently being used for fuel loading pattern optimization (LPO) problems. In this paper we have applied Estimation of distribution algorithm (EDA) to optimize initial core loading pattern (LP) of AHWR-LEU. In EDA, new solutions are generated by sampling the probability distribution model estimated from the selected best candidate solutions. The weighing factor ‘α’ decides the fraction of current best solution for updating the probability distribution function after each generation. A wider use of EDA warrants a comprehensive study on parameters like population size, weighing factor ‘α’ and initial probability distribution function. In the present study, we have done an extensive analysis on these parameters (population size, weighing factor ‘α’ and initial probability distribution function) in EDA. It is observed that choosing a very small value of ‘α’ may limit the search of optimized solutions in the near vicinity of initial probability distribution function and better loading patterns which are away from initial distribution function may not be considered with due weightage. It is also observed that increasing the population size improves the optimized loading pattern, however the algorithm still fails if the initial distribution function is not close to the expected optimized solution. We have tried to find out the suitable values for ‘α’ and population size to be considered for AHWR-LEU initial core loading pattern optimization problem. For sake of comparison and completeness, we have also addressed the
A practical algorithm for distribution state estimation including renewable energy sources
Energy Technology Data Exchange (ETDEWEB)
Niknam, Taher [Electronic and Electrical Department, Shiraz University of Technology, Modares Blvd., P.O. 71555-313, Shiraz (Iran); Firouzi, Bahman Bahmani [Islamic Azad University Marvdasht Branch, Marvdasht (Iran)
2009-11-15
Renewable energy is energy that is in continuous supply over time. These kinds of energy sources are divided into five principal renewable sources of energy: the sun, the wind, flowing water, biomass and heat from within the earth. According to some studies carried out by the research institutes, about 25% of the new generation will be generated by Renewable Energy Sources (RESs) in the near future. Therefore, it is necessary to study the impact of RESs on the power systems, especially on the distribution networks. This paper presents a practical Distribution State Estimation (DSE) including RESs and some practical consideration. The proposed algorithm is based on the combination of Nelder-Mead simplex search and Particle Swarm Optimization (PSO) algorithms, called PSO-NM. The proposed algorithm can estimate load and RES output values by Weighted Least-Square (WLS) approach. Some practical considerations are var compensators, Voltage Regulators (VRs), Under Load Tap Changer (ULTC) transformer modeling, which usually have nonlinear and discrete characteristics, and unbalanced three-phase power flow equations. The comparison results with other evolutionary optimization algorithms such as original PSO, Honey Bee Mating Optimization (HBMO), Neural Networks (NNs), Ant Colony Optimization (ACO), and Genetic Algorithm (GA) for a test system demonstrate that PSO-NM is extremely effective and efficient for the DSE problems. (author)
Energy Technology Data Exchange (ETDEWEB)
Lober, R.R.; Tautges, T.J.; Vaughan, C.T.
1997-03-01
Paving is an automated mesh generation algorithm which produces all-quadrilateral elements. It can additionally generate these elements in varying sizes such that the resulting mesh adapts to a function distribution, such as an error function. While powerful, conventional paving is a very serial algorithm in its operation. Parallel paving is the extension of serial paving into parallel environments to perform the same meshing functions as conventional paving only on distributed, discretized models. This extension allows large, adaptive, parallel finite element simulations to take advantage of paving`s meshing capabilities for h-remap remeshing. A significantly modified version of the CUBIT mesh generation code has been developed to host the parallel paving algorithm and demonstrate its capabilities on both two dimensional and three dimensional surface geometries and compare the resulting parallel produced meshes to conventionally paved meshes for mesh quality and algorithm performance. Sandia`s {open_quotes}tiling{close_quotes} dynamic load balancing code has also been extended to work with the paving algorithm to retain parallel efficiency as subdomains undergo iterative mesh refinement.
Directory of Open Access Journals (Sweden)
Kazem Mohammadi- Aghdam
2015-10-01
Full Text Available This paper proposes the application of a new version of the heuristic particle swarm optimization (PSO method for designing water distribution networks (WDNs. The optimization problem of looped water distribution networks is recognized as an NP-hard combinatorial problem which cannot be easily solved using traditional mathematical optimization techniques. In this paper, the concept of dynamic swarm size is considered in an attempt to increase the convergence speed of the original PSO algorithm. In this strategy, the size of the swarm is dynamically changed according to the iteration number of the algorithm. Furthermore, a novel mutation approach is introduced to increase the diversification property of the PSO and to help the algorithm to avoid trapping in local optima. The new version of the PSO algorithm is called dynamic mutated particle swarm optimization (DMPSO. The proposed DMPSO is then applied to solve WDN design problems. Finally, two illustrative examples are used for comparison to verify the efficiency of the proposed DMPSO as compared to other intelligent algorithms.
Directory of Open Access Journals (Sweden)
Hui He
2013-01-01
Full Text Available It is of great significance to research the early warning system for large-scale network security incidents. It can improve the network system’s emergency response capabilities, alleviate the cyber attacks’ damage, and strengthen the system’s counterattack ability. A comprehensive early warning system is presented in this paper, which combines active measurement and anomaly detection. The key visualization algorithm and technology of the system are mainly discussed. The large-scale network system’s plane visualization is realized based on the divide and conquer thought. First, the topology of the large-scale network is divided into some small-scale networks by the MLkP/CR algorithm. Second, the sub graph plane visualization algorithm is applied to each small-scale network. Finally, the small-scale networks’ topologies are combined into a topology based on the automatic distribution algorithm of force analysis. As the algorithm transforms the large-scale network topology plane visualization problem into a series of small-scale network topology plane visualization and distribution problems, it has higher parallelism and is able to handle the display of ultra-large-scale network topology.
An Extended Genetic Algorithm for Distributed Integration of Fuzzy Process Planning and Scheduling
Directory of Open Access Journals (Sweden)
Shuai Zhang
2016-01-01
Full Text Available The distributed integration of process planning and scheduling (DIPPS aims to simultaneously arrange the two most important manufacturing stages, process planning and scheduling, in a distributed manufacturing environment. Meanwhile, considering its advantage corresponding to actual situation, the triangle fuzzy number (TFN is adopted in DIPPS to represent the machine processing and transportation time. In order to solve this problem and obtain the optimal or near-optimal solution, an extended genetic algorithm (EGA with innovative three-class encoding method, improved crossover, and mutation strategies is proposed. Furthermore, a local enhancement strategy featuring machine replacement and order exchange is also added to strengthen the local search capability on the basic process of genetic algorithm. Through the verification of experiment, EGA achieves satisfactory results all in a very short period of time and demonstrates its powerful performance in dealing with the distributed integration of fuzzy process planning and scheduling (DIFPPS.
A Reputation-based Distributed District Scheduling Algorithm for Smart Grids
Directory of Open Access Journals (Sweden)
D. Borra
2015-05-01
Full Text Available In this paper we develop and test a distributed algorithm providing Energy Consumption Schedules (ECS in smart grids for a residential district. The goal is to achieve a given aggregate load prole. The NP-hard constrained optimization problem reduces to a distributed unconstrained formulation by means of Lagrangian Relaxation technique, and a meta-heuristic algorithm based on a Quantum inspired Particle Swarm with Levy flights. A centralized iterative reputation-reward mechanism is proposed for end-users to cooperate to avoid power peaks and reduce global overload, based on random distributions simulating human behaviors and penalties on the eective ECS diering from the suggested ECS. Numerical results show the protocols eectiveness.
An Efficient Distributed Algorithm for Constructing Spanning Trees in Wireless Sensor Networks
Directory of Open Access Journals (Sweden)
Rosana Lachowski
2015-01-01
Full Text Available Monitoring and data collection are the two main functions in wireless sensor networks (WSNs. Collected data are generally transmitted via multihop communication to a special node, called the sink. While in a typical WSN, nodes have a sink node as the final destination for the data traffic, in an ad hoc network, nodes need to communicate with each other. For this reason, routing protocols for ad hoc networks are inefficient for WSNs. Trees, on the other hand, are classic routing structures explicitly or implicitly used in WSNs. In this work, we implement and evaluate distributed algorithms for constructing routing trees in WSNs described in the literature. After identifying the drawbacks and advantages of these algorithms, we propose a new algorithm for constructing spanning trees in WSNs. The performance of the proposed algorithm and the quality of the constructed tree were evaluated in different network scenarios. The results showed that the proposed algorithm is a more efficient solution. Furthermore, the algorithm provides multiple routes to the sensor nodes to be used as mechanisms for fault tolerance and load balancing.
A hybrid genetic algorithm for the distributed permutation flowshop scheduling problem
Directory of Open Access Journals (Sweden)
Jian Gao
2011-08-01
Full Text Available Distributed Permutation Flowshop Scheduling Problem (DPFSP is a newly proposed scheduling problem, which is a generalization of classical permutation flow shop scheduling problem. The DPFSP is NP-hard in general. It is in the early stages of studies on algorithms for solving this problem. In this paper, we propose a GA-based algorithm, denoted by GA_LS, for solving this problem with objective to minimize the maximum completion time. In the proposed GA_LS, crossover and mutation operators are designed to make it suitable for the representation of DPFSP solutions, where the set of partial job sequences is employed. Furthermore, GA_LS utilizes an efficient local search method to explore neighboring solutions. The local search method uses three proposed rules that move jobs within a factory or between two factories. Intensive experiments on the benchmark instances, extended from Taillard instances, are carried out. The results indicate that the proposed hybrid genetic algorithm can obtain better solutions than all the existing algorithms for the DPFSP, since it obtains better relative percentage deviation and differences of the results are also statistically significant. It is also seen that best-known solutions for most instances are updated by our algorithm. Moreover, we also show the efficiency of the GA_LS by comparing with similar genetic algorithms with the existing local search methods.
Multilevel sequential Monte-Carlo samplers
Jasra, Ajay
2016-01-01
Multilevel Monte-Carlo methods provide a powerful computational technique for reducing the computational cost of estimating expectations for a given computational effort. They are particularly relevant for computational problems when approximate distributions are determined via a resolution parameter h, with h=0 giving the theoretical exact distribution (e.g. SDEs or inverse problems with PDEs). The method provides a benefit by coupling samples from successive resolutions, and estimating differences of successive expectations. We develop a methodology that brings Sequential Monte-Carlo (SMC) algorithms within the framework of the Multilevel idea, as SMC provides a natural set-up for coupling samples over different resolutions. We prove that the new algorithm indeed preserves the benefits of the multilevel principle, even if samples at all resolutions are now correlated.
Multilevel sequential Monte-Carlo samplers
Jasra, Ajay
2016-01-05
Multilevel Monte-Carlo methods provide a powerful computational technique for reducing the computational cost of estimating expectations for a given computational effort. They are particularly relevant for computational problems when approximate distributions are determined via a resolution parameter h, with h=0 giving the theoretical exact distribution (e.g. SDEs or inverse problems with PDEs). The method provides a benefit by coupling samples from successive resolutions, and estimating differences of successive expectations. We develop a methodology that brings Sequential Monte-Carlo (SMC) algorithms within the framework of the Multilevel idea, as SMC provides a natural set-up for coupling samples over different resolutions. We prove that the new algorithm indeed preserves the benefits of the multilevel principle, even if samples at all resolutions are now correlated.
DEFF Research Database (Denmark)
Tonelli, Oscar; Berardinelli, Gilberto; Tavares, Fernando Menezes Leitão
2013-01-01
Next generation wireless networks aim at a significant improvement of the spectral efficiency in order to meet the dramatic increase in data service demand. In local area scenarios user-deployed base stations are expected to take place, thus making the centralized planning of frequency resources...... activities with the Autonomous Component Carrier Selection (ACCS) algorithm, a distributed solution for interference management among small neighboring cells. A preliminary evaluation of the algorithm performance is provided considering its live execution on a software defined radio network testbed...
Multi-agent coordination algorithms for control of distributed energy resources in smart grids
Cortes, Andres
Sustainable energy is a top-priority for researchers these days, since electricity and transportation are pillars of modern society. Integration of clean energy technologies such as wind, solar, and plug-in electric vehicles (PEVs), is a major engineering challenge in operation and management of power systems. This is due to the uncertain nature of renewable energy technologies and the large amount of extra load that PEVs would add to the power grid. Given the networked structure of a power system, multi-agent control and optimization strategies are natural approaches to address the various problems of interest for the safe and reliable operation of the power grid. The distributed computation in multi-agent algorithms addresses three problems at the same time: i) it allows for the handling of problems with millions of variables that a single processor cannot compute, ii) it allows certain independence and privacy to electricity customers by not requiring any usage information, and iii) it is robust to localized failures in the communication network, being able to solve problems by simply neglecting the failing section of the system. We propose various algorithms to coordinate storage, generation, and demand resources in a power grid using multi-agent computation and decentralized decision making. First, we introduce a hierarchical vehicle-one-grid (V1G) algorithm for coordination of PEVs under usage constraints, where energy only flows from the grid in to the batteries of PEVs. We then present a hierarchical vehicle-to-grid (V2G) algorithm for PEV coordination that takes into consideration line capacity constraints in the distribution grid, and where energy flows both ways, from the grid in to the batteries, and from the batteries to the grid. Next, we develop a greedy-like hierarchical algorithm for management of demand response events with on/off loads. Finally, we introduce distributed algorithms for the optimal control of distributed energy resources, i
A distributed multi-agent linear biobjective algorithm for energy flow optimization in microgrids
DEFF Research Database (Denmark)
Brehm, Robert; Top, Søren; Mátéfi-Tempfli, Stefan
2016-01-01
consisting of local energy resources and storage capacities is presented which is based on the auction algorithm for assignment problems originally introduced by Bertsekas in 1979 [1]. It is shown that the topology of a microgrid can be represented as a bipartite graph and mathematically be described...... as a classical transportation problem. This allows applying an auction algorithm scheme in a distributed way where each energy supply system node is either a source or a sink and is represented by an individual acting agent. The single-objective approach is extended towards bi-objectivity to build a framework...
Energy Technology Data Exchange (ETDEWEB)
Schmidt, Daniel, E-mail: frank.wiekhorst@ptb.de; Eberbeck, Dietmar; Steinhoff, Uwe; Wiekhorst, Frank
2017-06-01
The characterization of the size distribution of magnetic nanoparticles is an important step for the evaluation of their suitability for many different applications like magnetic hyperthermia, drug targeting or Magnetic Particle Imaging. We present a new method based on the iterative Kaczmarz algorithm that enables the reconstruction of the size distribution from magnetization measurements without a priori knowledge of the distribution form. We show in simulations that the method is capable of very exact reconstructions of a given size distribution and, in that, is highly robust to noise contamination. Moreover, we applied the method on the well characterized FeraSpin™ series and obtained results that were in accordance with literature and boundary conditions based on their synthesis via separation of the original suspension FeraSpin R. It is therefore concluded that this method is a powerful and intuitive tool for reconstructing particle size distributions from magnetization measurements. - Highlights: • A new method for the size distribution fit of magnetic nanoparticles is proposed. • Employed Kaczmarz algorithm does not need a priori input or eigenwert regularization. • The method is highly robust to noise contamination. • Size distributions are reconstructed from simulated and measured magnetization curves.
Supply chain management using fp-growth algorithm for medicine distribution
Wahana, A.; Maylawati, D. S.; Irfan, M.; Effendy, H.
2018-03-01
Distribution of drugs evenly in accordance with the needs of Public Health Center (Puskesmas) become one of the responsibilities by the Health Office in Indonesia. This study aims to provide recommendations for distribution of drugs from the Department of Health according to the needs of each Puskesmas. Because often the distribution of drugs is not in accordance with the needs of medicines stock of each Puskesmas. This causes the possibility of drug stock void in Puskesmas in need, while there is excess stock of drugs in Puskesmas that do not need. Supply Chain Management (SCM) is applied as a controlling drug stock at Puskesmas in order to avoid drug vacuum or excess drug that eventually unused. In addition, the Frequent Pattern Growth (FP-Growth) algorithm that generates frequent item sets is used to provide drug distribution recommendations by looking at the highest frequency of drug occurrences and frequent drug frequencies. Based on testing black box system conducted in Health Office Purwakarta regency of Indonesia which oversees 20 Puskesmas with 100 data of drug distribution transactions, it can be concluded that system functionality is running well and SCM successfully implemented to arrange distribution process of medicine well. Furthermore FP-Growth algorithm was able to provide recommendations for distribution of drugs with a high success rate. This is evidenced by the test results with various combinations of input parameters, FP-Growth is able to produce the right frequent item sets.
International Nuclear Information System (INIS)
Santamaria-Fernandez, Rebeca; Cave, Mark R.; Hill, Steve J.
2006-01-01
A study of the trace metal distribution in sediment samples from the Galician coast (Spain) has been performed. A multielement extraction method optimised via experimental design has been employed. The method uses centrifugation to pass the extractant solution at varying pH, through the sediment sample. The sequential leaches were collected and analysed by ICP-AES. Chemometric approaches were utilised to identify the composition of the physico-chemical components in order to characterise the sample. The samples collected at different sites could be classified according to their differences in metal bio-availability and important information regarding element distribution within the physico-chemical components is given. The method has proved to be a quick and reliable way to evaluate sediment samples for environmental geochemistry analysis. In addition, this approach has potential as fast screening method for the bio-availability of metals in the environment
Ogawa, Masakatsu; Hiraguri, Takefumi; Nishimori, Kentaro; Takaya, Kazuhiro; Murakawa, Kazuo
This paper proposes and investigates a distributed adaptive contention window adjustment algorithm based on the transmission history for wireless LANs called the transmission-history-based distributed adaptive contention window adjustment (THAW) algorithm. The objective of this paper is to reduce the transmission delay and improve the channel throughput compared to conventional algorithms. The feature of THAW is that it adaptively adjusts the initial contention window (CWinit) size in the binary exponential backoff (BEB) algorithm used in the IEEE 802.11 standard according to the transmission history and the automatic rate fallback (ARF) algorithm, which is the most basic algorithm in automatic rate controls. This effect is to keep CWinit at a high value in a congested state. Simulation results show that the THAW algorithm outperforms the conventional algorithms in terms of the channel throughput and delay, even if the timer in the ARF is changed.
Wang, Qianqian; Zhao, Jing; Gong, Yong; Hao, Qun; Peng, Zhong
2017-11-20
A hybrid artificial bee colony (ABC) algorithm inspired by the best-so-far solution and bacterial chemotaxis was introduced to optimize the parameters of the five-parameter bidirectional reflectance distribution function (BRDF) model. To verify the performance of the hybrid ABC algorithm, we measured BRDF of three kinds of samples and simulated the undetermined parameters of the five-parameter BRDF model using the hybrid ABC algorithm and the genetic algorithm, respectively. The experimental results demonstrate that the hybrid ABC algorithm outperforms the genetic algorithm in convergence speed, accuracy, and time efficiency under the same conditions.
Improved Cost-Base Design of Water Distribution Networks using Genetic Algorithm
Moradzadeh Azar, Foad; Abghari, Hirad; Taghi Alami, Mohammad; Weijs, Steven
2010-05-01
Population growth and progressive extension of urbanization in different places of Iran cause an increasing demand for primary needs. The water, this vital liquid is the most important natural need for human life. Providing this natural need is requires the design and construction of water distribution networks, that incur enormous costs on the country's budget. Any reduction in these costs enable more people from society to access extreme profit least cost. Therefore, investment of Municipal councils need to maximize benefits or minimize expenditures. To achieve this purpose, the engineering design depends on the cost optimization techniques. This paper, presents optimization models based on genetic algorithm(GA) to find out the minimum design cost Mahabad City's (North West, Iran) water distribution network. By designing two models and comparing the resulting costs, the abilities of GA were determined. the GA based model could find optimum pipe diameters to reduce the design costs of network. Results show that the water distribution network design using Genetic Algorithm could lead to reduction of at least 7% in project costs in comparison to the classic model. Keywords: Genetic Algorithm, Optimum Design of Water Distribution Network, Mahabad City, Iran.
International Nuclear Information System (INIS)
Castellini, P; Cecchini, S; Stroppa, L; Paone, N
2015-01-01
The paper presents an adaptive illumination system for image quality enhancement in vision-based quality control systems. In particular, a spatial modulation of illumination intensity is proposed in order to improve image quality, thus compensating for different target scattering properties, local reflections and fluctuations of ambient light. The desired spatial modulation of illumination is obtained by a digital light projector, used to illuminate the scene with an arbitrary spatial distribution of light intensity, designed to improve feature extraction in the region of interest. The spatial distribution of illumination is optimized by running a genetic algorithm. An image quality estimator is used to close the feedback loop and to stop iterations once the desired image quality is reached. The technique proves particularly valuable for optimizing the spatial illumination distribution in the region of interest, with the remarkable capability of the genetic algorithm to adapt the light distribution to very different target reflectivity and ambient conditions. The final objective of the proposed technique is the improvement of the matching score in the recognition of parts through matching algorithms, hence of the diagnosis of machine vision-based quality inspections. The procedure has been validated both by a numerical model and by an experimental test, referring to a significant problem of quality control for the washing machine manufacturing industry: the recognition of a metallic clamp. Its applicability to other domains is also presented, specifically for the visual inspection of shoes with retro-reflective tape and T-shirts with paillettes. (paper)
Indian Academy of Sciences (India)
algorithm design technique called 'divide-and-conquer'. One of ... Turtle graphics, September. 1996. 5. ... whole list named 'PO' is a pointer to the first element of the list; ..... Program for computing matrices X and Y and placing the result in C *).
Indian Academy of Sciences (India)
algorithm that it is implicitly understood that we know how to generate the next natural ..... Explicit comparisons are made in line (1) where maximum and minimum is ... It can be shown that the function T(n) = 3/2n -2 is the solution to the above ...
Directory of Open Access Journals (Sweden)
Young-Jae Song
2009-07-01
Full Text Available Existing anomaly and intrusion detection schemes of wireless sensor networks have mainly focused on the detection of intrusions. Once the intrusion is detected, an alerts or claims will be generated. However, any unidentified malicious nodes in the network could send faulty anomaly and intrusion claims about the legitimate nodes to the other nodes. Verifying the validity of such claims is a critical and challenging issue that is not considered in the existing cooperative-based distributed anomaly and intrusion detection schemes of wireless sensor networks. In this paper, we propose a validation algorithm that addresses this problem. This algorithm utilizes the concept of intrusion-aware reliability that helps to provide adequate reliability at a modest communication cost. In this paper, we also provide a security resiliency analysis of the proposed intrusion-aware alert validation algorithm.
Research on fully distributed optical fiber sensing security system localization algorithm
Wu, Xu; Hou, Jiacheng; Liu, Kun; Liu, Tiegen
2013-12-01
A new fully distributed optical fiber sensing and location technology based on the Mach-Zehnder interferometers is studied. In this security system, a new climbing point locating algorithm based on short-time average zero-crossing rate is presented. By calculating the zero-crossing rates of the multiple grouped data separately, it not only utilizes the advantages of the frequency analysis method to determine the most effective data group more accurately, but also meets the requirement of the real-time monitoring system. Supplemented with short-term energy calculation group signal, the most effective data group can be quickly picked out. Finally, the accurate location of the climbing point can be effectively achieved through the cross-correlation localization algorithm. The experimental results show that the proposed algorithm can realize the accurate location of the climbing point and meanwhile the outside interference noise of the non-climbing behavior can be effectively filtered out.
He, Zhenzong; Qi, Hong; Wang, Yuqing; Ruan, Liming
2014-10-01
Four improved Ant Colony Optimization (ACO) algorithms, i.e. the probability density function based ACO (PDF-ACO) algorithm, the Region ACO (RACO) algorithm, Stochastic ACO (SACO) algorithm and Homogeneous ACO (HACO) algorithm, are employed to estimate the particle size distribution (PSD) of the spheroidal particles. The direct problems are solved by the extended Anomalous Diffraction Approximation (ADA) and the Lambert-Beer law. Three commonly used monomodal distribution functions i.e. the Rosin-Rammer (R-R) distribution function, the normal (N-N) distribution function, and the logarithmic normal (L-N) distribution function are estimated under dependent model. The influence of random measurement errors on the inverse results is also investigated. All the results reveal that the PDF-ACO algorithm is more accurate than the other three ACO algorithms and can be used as an effective technique to investigate the PSD of the spheroidal particles. Furthermore, the Johnson's SB (J-SB) function and the modified beta (M-β) function are employed as the general distribution functions to retrieve the PSD of spheroidal particles using PDF-ACO algorithm. The investigation shows a reasonable agreement between the original distribution function and the general distribution function when only considering the variety of the length of the rotational semi-axis.
International Nuclear Information System (INIS)
Zhou Shumin; Sun Yamin; Tang Bin
2007-01-01
In order to enhance the time synchronization quality of the distributed system, a time synchronization algorithm of distributed system based on server time-revise and workstation self-adjust is proposed. The time-revise cycle and self-adjust process is introduced in the paper. The algorithm reduces network flow effectively and enhances the quality of clock-synchronization. (authors)
Energy Technology Data Exchange (ETDEWEB)
Niknam, Taher; Meymand, Hamed Zeinoddini; Nayeripour, Majid [Electrical and Electronic Engineering Department, Shiraz University of Technology, Shiraz (Iran)
2010-08-15
Fuel cell power plants (FCPPs) have been taken into a great deal of consideration in recent years. The continuing growth of the power demand together with environmental constraints is increasing interest to use FCPPs in power system. Since FCPPs are usually connected to distribution network, the effect of FCPPs on distribution network is more than other sections of power system. One of the most important issues in distribution networks is optimal operation management (OOM) which can be affected by FCPPs. This paper proposes a new approach for optimal operation management of distribution networks including FCCPs. In the article, we consider the total electrical energy losses, the total electrical energy cost and the total emission as the objective functions which should be minimized. Whereas the optimal operation in distribution networks has a nonlinear mixed integer optimization problem, the optimal solution could be obtained through an evolutionary method. We use a new evolutionary algorithm based on Fuzzy Adaptive Particle Swarm Optimization (FAPSO) to solve the optimal operation problem and compare this method with Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Differential Evolution (DE), Ant Colony Optimization (ACO) and Tabu Search (TS) over two distribution test feeders. (author)
Network reconfiguration for loss reduction in electrical distribution system using genetic algorithm
International Nuclear Information System (INIS)
Adail, A.S.A.A.
2012-01-01
Distribution system is a critical links between the utility and the nuclear installation. During feeding electricity to that installation there are power losses. The quality of the network depends on the reduction of these losses. Distribution system which feeds the nuclear installation must have a higher quality power. For example, in Inshas site, electrical power is supplied to it from two incoming feeders (one from new abu-zabal substation and the other from old abu-zabal substation). Each feeder is designed to carry the full load, while the operator preferred to connect with a new abu-zabal substation, which has a good power quality. Bad power quality affects directly the nuclear reactor and has a negative impact on the installed sensitive equipment's of the operation. The thesis is Studying the electrical losses in a distribution system (causes and effected factors), feeder reconfiguration methods, and applying of genetic algorithm in an electric distribution power system. In the end, this study proposes an optimization technique based on genetic algorithms for distribution network reconfiguration to reduce the network losses to minimum. The proposed method is applied to IEEE test network; that contain 3 feeders and 16 nodes. The technique is applied through two groups, distribution have general loads, and nuclear loads. In the groups the technique applied to seven cases at normal operation state, system fault condition as well as different loads conditions. Simulated results are drawn to show the accuracy of the technique.
Bizer, David S; DeMarzo, Peter M
1992-01-01
The authors study environments in which agents may borrow sequentially from more than one leader. Although debt is prioritized, additional lending imposes an externality on prior debt because, with moral hazard, the probability of repayment of prior loans decreases. Equilibrium interest rates are higher than they would be if borrowers could commit to borrow from at most one bank. Even though the loan terms are less favorable than they would be under commitment, the indebtedness of borrowers i...
Dong, Yu-Shuang; Xu, Gao-Chao; Fu, Xiao-Dong
2014-01-01
The cloud platform provides various services to users. More and more cloud centers provide infrastructure as the main way of operating. To improve the utilization rate of the cloud center and to decrease the operating cost, the cloud center provides services according to requirements of users by sharding the resources with virtualization. Considering both QoS for users and cost saving for cloud computing providers, we try to maximize performance and minimize energy cost as well. In this paper, we propose a distributed parallel genetic algorithm (DPGA) of placement strategy for virtual machines deployment on cloud platform. It executes the genetic algorithm parallelly and distributedly on several selected physical hosts in the first stage. Then it continues to execute the genetic algorithm of the second stage with solutions obtained from the first stage as the initial population. The solution calculated by the genetic algorithm of the second stage is the optimal one of the proposed approach. The experimental results show that the proposed placement strategy of VM deployment can ensure QoS for users and it is more effective and more energy efficient than other placement strategies on the cloud platform.
Directory of Open Access Journals (Sweden)
Yu-Shuang Dong
2014-01-01
Full Text Available The cloud platform provides various services to users. More and more cloud centers provide infrastructure as the main way of operating. To improve the utilization rate of the cloud center and to decrease the operating cost, the cloud center provides services according to requirements of users by sharding the resources with virtualization. Considering both QoS for users and cost saving for cloud computing providers, we try to maximize performance and minimize energy cost as well. In this paper, we propose a distributed parallel genetic algorithm (DPGA of placement strategy for virtual machines deployment on cloud platform. It executes the genetic algorithm parallelly and distributedly on several selected physical hosts in the first stage. Then it continues to execute the genetic algorithm of the second stage with solutions obtained from the first stage as the initial population. The solution calculated by the genetic algorithm of the second stage is the optimal one of the proposed approach. The experimental results show that the proposed placement strategy of VM deployment can ensure QoS for users and it is more effective and more energy efficient than other placement strategies on the cloud platform.
Implementation of an evolutionary algorithm in planning investment in a power distribution system
Directory of Open Access Journals (Sweden)
Carlos Andrés García Montoya
2011-06-01
Full Text Available The definition of an investment plan to implement in a distribution power system, is a task that constantly faced by utilities. This work presents a methodology for determining the investment plan for a distribution power system under a shortterm, using as a criterion for evaluating investment projects, associated costs and customers benefit from its implementation. Given the number of projects carried out annually on the system, the definition of an investment plan requires the use of computational tools to evaluate, a set of possibilities, the one that best suits the needs of the present system and better results. That is why in the job, implementing a multi objective evolutionary algorithm SPEA (Strength Pareto Evolutionary Algorithm, which, based on the principles of Pareto optimality, it deliver to the planning expert, the best solutions found in the optimization process. The performance of the algorithm is tested using a set of projects to determine the best among the possible plans. We analyze also the effect of operators on the performance of evolutionary algorithm and results.
Directory of Open Access Journals (Sweden)
Junlong Zhu
2017-01-01
Full Text Available We consider a distributed constrained optimization problem over a time-varying network, where each agent only knows its own cost functions and its constraint set. However, the local constraint set may not be known in advance or consists of huge number of components in some applications. To deal with such cases, we propose a distributed stochastic subgradient algorithm over time-varying networks, where the estimate of each agent projects onto its constraint set by using random projection technique and the implement of information exchange between agents by employing asynchronous broadcast communication protocol. We show that our proposed algorithm is convergent with probability 1 by choosing suitable learning rate. For constant learning rate, we obtain an error bound, which is defined as the expected distance between the estimates of agent and the optimal solution. We also establish an asymptotic upper bound between the global objective function value at the average of the estimates and the optimal value.
Multilayer PV-storage Microgrids Algorithm for the Dispatch of Distributed Network
Directory of Open Access Journals (Sweden)
Yang Ping
2016-01-01
Full Text Available In recent years, due to the support of our country, PV-storage microgrid develops rapidly. However, the flexible network operation modes of PV-storage microgrid change flexibly and the operating characteristics with a large amout of sources is highly complicated. Based on the existing microgrid coordinate control methods, this paper proposes multilayer PV-storage microgrid algorithm for fitting dispatch of distributed network, which achieves maximum output of renewable energy when meeting the scheduling requirements of network, by building PV-storage microgrid type dynamic simulation system in a variety of conditions in PSCAD. Simulation results show that the heuristic algorithm proposed can achieve microgrid stable operation and satisfy the demands of the dispatch in distributed network.
Optimal Location and Sizing of UPQC in Distribution Networks Using Differential Evolution Algorithm
Directory of Open Access Journals (Sweden)
Seyed Abbas Taher
2012-01-01
Full Text Available Differential evolution (DE algorithm is used to determine optimal location of unified power quality conditioner (UPQC considering its size in the radial distribution systems. The problem is formulated to find the optimum location of UPQC based on an objective function (OF defined for improving of voltage and current profiles, reducing power loss and minimizing the investment costs considering the OF's weighting factors. Hence, a steady-state model of UPQC is derived to set in forward/backward sweep load flow. Studies are performed on two IEEE 33-bus and 69-bus standard distribution networks. Accuracy was evaluated by reapplying the procedures using both genetic (GA and immune algorithms (IA. Comparative results indicate that DE is capable of offering a nearer global optimal in minimizing the OF and reaching all the desired conditions than GA and IA.
Mwakabuta, Ndaga Stanslaus
Electric power distribution systems play a significant role in providing continuous and "quality" electrical energy to different classes of customers. In the context of the present restrictions on transmission system expansions and the new paradigm of "open and shared" infrastructure, new approaches to distribution system analyses, economic and operational decision-making need investigation. This dissertation includes three layers of distribution system investigations. In the basic level, improved linear models are shown to offer significant advantages over previous models for advanced analysis. In the intermediate level, the improved model is applied to solve the traditional problem of operating cost minimization using capacitors and voltage regulators. In the advanced level, an artificial intelligence technique is applied to minimize cost under Distributed Generation injection from private vendors. Soft computing techniques are finding increasing applications in solving optimization problems in large and complex practical systems. The dissertation focuses on Genetic Algorithm for investigating the economic aspects of distributed generation penetration without compromising the operational security of the distribution system. The work presents a methodology for determining the optimal pricing of distributed generation that would help utilities make a decision on how to operate their system economically. This would enable modular and flexible investments that have real benefits to the electric distribution system. Improved reliability for both customers and the distribution system in general, reduced environmental impacts, increased efficiency of energy use, and reduced costs of energy services are some advantages.
Indian Academy of Sciences (India)
will become clear in the next article when we discuss a simple logo like programming language. ... Rod B may be used as an auxiliary store. The problem is to find an algorithm which performs this task. ... No disks are moved from A to Busing C as auxiliary rod. • move _disk (A, C);. (No + l)th disk is moved from A to C directly ...
Distributed Energy-Efficient Topology Control Algorithm in Home M2M Networks
Lee, Chao-Yang; Yang, Chu-Sing
2012-01-01
Because machine-to-machine (M2M) technology enables machines to communicate with each other without human intervention, it could play a big role in sensor network systems. Through wireless sensor network (WSN) gateways, various information can be collected by sensors for M2M systems. For home M2M networks, this study proposes a distributed energy-efficient topology control algorithm for both topology construction and topology maintenance. Topology control is an effective method of enhancing e...
AbdulJabbar, Mustafa Abdulmajeed
2017-05-11
Reduction of communication and efficient partitioning are key issues for achieving scalability in hierarchical N-Body algorithms like Fast Multipole Method (FMM). In the present work, we propose three independent strategies to improve partitioning and reduce communication. First, we show that the conventional wisdom of using space-filling curve partitioning may not work well for boundary integral problems, which constitute a significant portion of FMM’s application user base. We propose an alternative method that modifies orthogonal recursive bisection to relieve the cell-partition misalignment that has kept it from scaling previously. Secondly, we optimize the granularity of communication to find the optimal balance between a bulk-synchronous collective communication of the local essential tree and an RDMA per task per cell. Finally, we take the dynamic sparse data exchange proposed by Hoefler et al. [1] and extend it to a hierarchical sparse data exchange, which is demonstrated at scale to be faster than the MPI library’s MPI_Alltoallv that is commonly used.
Impelluso, Thomas J
2003-06-01
An algorithm for bone remodeling is presented which allows for both a redistribution of density and a continuous change of principal material directions for the orthotropic material properties of bone. It employs a modal analysis to add density for growth and a local effective strain based analysis to redistribute density. General re-distribution functions are presented. The model utilizes theories of cellular solids to relate density and strength. The code predicts the same general density distributions and local orthotropy as observed in reality.
Energy Technology Data Exchange (ETDEWEB)
Hong, Y.-Y.; Chen, Y.-C. [Chung Yuan University (China). Dept. of Electrical Engineering
1999-05-01
A new method is proposed for locating multiple harmonic sources in distribution systems. The proposed method first determines the proper locations for metering measurement using fuzzy clustering. Next, an artificial neural network based on the back-propagation approach is used to identify the most likely location for multiple harmonic sources. A set of systematic algorithmic steps is developed until all harmonic locations are identified. The simulation results for an 18-busbar system show that the proposed method is very efficient in locating the multiple harmonics in a distribution system. (author)
Sanyal, Soumya; Jain, Amit; Das, Sajal K.; Biswas, Rupak
2003-01-01
In this paper, we propose a distributed approach for mapping a single large application to a heterogeneous grid environment. To minimize the execution time of the parallel application, we distribute the mapping overhead to the available nodes of the grid. This approach not only provides a fast mapping of tasks to resources but is also scalable. We adopt a hierarchical grid model and accomplish the job of mapping tasks to this topology using a scheduler tree. Results show that our three-phase algorithm provides high quality mappings, and is fast and scalable.
WANG, Qingrong; ZHU, Changfeng
2017-06-01
Integration of distributed heterogeneous data sources is the key issues under the big data applications. In this paper the strategy of variable precision is introduced to the concept lattice, and the one-to-one mapping mode of variable precision concept lattice and ontology concept lattice is constructed to produce the local ontology by constructing the variable precision concept lattice for each subsystem, and the distributed generation algorithm of variable precision concept lattice based on ontology heterogeneous database is proposed to draw support from the special relationship between concept lattice and ontology construction. Finally, based on the standard of main concept lattice of the existing heterogeneous database generated, a case study has been carried out in order to testify the feasibility and validity of this algorithm, and the differences between the main concept lattice and the standard concept lattice are compared. Analysis results show that this algorithm above-mentioned can automatically process the construction process of distributed concept lattice under the heterogeneous data sources.
Zheng, Yan
2015-03-01
Internet of things (IoT), focusing on providing users with information exchange and intelligent control, attracts a lot of attention of researchers from all over the world since the beginning of this century. IoT is consisted of large scale of sensor nodes and data processing units, and the most important features of IoT can be illustrated as energy confinement, efficient communication and high redundancy. With the sensor nodes increment, the communication efficiency and the available communication band width become bottle necks. Many research work is based on the instance which the number of joins is less. However, it is not proper to the increasing multi-join query in whole internet of things. To improve the communication efficiency between parallel units in the distributed sensor network, this paper proposed parallel query optimization algorithm based on distribution attributes cost graph. The storage information relations and the network communication cost are considered in this algorithm, and an optimized information changing rule is established. The experimental result shows that the algorithm has good performance, and it would effectively use the resource of each node in the distributed sensor network. Therefore, executive efficiency of multi-join query between different nodes could be improved.
Fuzzy-Logic Based Distributed Energy-Efficient Clustering Algorithm for Wireless Sensor Networks.
Zhang, Ying; Wang, Jun; Han, Dezhi; Wu, Huafeng; Zhou, Rundong
2017-07-03
Due to the high-energy efficiency and scalability, the clustering routing algorithm has been widely used in wireless sensor networks (WSNs). In order to gather information more efficiently, each sensor node transmits data to its Cluster Head (CH) to which it belongs, by multi-hop communication. However, the multi-hop communication in the cluster brings the problem of excessive energy consumption of the relay nodes which are closer to the CH. These nodes' energy will be consumed more quickly than the farther nodes, which brings the negative influence on load balance for the whole networks. Therefore, we propose an energy-efficient distributed clustering algorithm based on fuzzy approach with non-uniform distribution (EEDCF). During CHs' election, we take nodes' energies, nodes' degree and neighbor nodes' residual energies into consideration as the input parameters. In addition, we take advantage of Takagi, Sugeno and Kang (TSK) fuzzy model instead of traditional method as our inference system to guarantee the quantitative analysis more reasonable. In our scheme, each sensor node calculates the probability of being as CH with the help of fuzzy inference system in a distributed way. The experimental results indicate EEDCF algorithm is better than some current representative methods in aspects of data transmission, energy consumption and lifetime of networks.
International Nuclear Information System (INIS)
Ramos Muñoz, Edgar; Razeghi, Ghazal; Zhang, Li; Jabbari, Faryar
2016-01-01
The need to reduce greenhouse gas emissions and fossil fuel consumption has increased the popularity of plug-in electric vehicles. However, a large penetration of plug-in electric vehicles can pose challenges at the grid and local distribution levels. Various charging strategies have been proposed to address such challenges, often separately. In this paper, it is shown that, with uncoordinated charging, distribution transformers and the grid can operate under highly undesirable conditions. Next, several strategies that require modest communication efforts are proposed to mitigate the burden created by high concentrations of plug-in electric vehicles, at the grid and local levels. Existing transformer and battery electric vehicle characteristics are used along with the National Household Travel Survey to simulate various charging strategies. It is shown through the analysis of hot spot temperature and equivalent aging factor that the coordinated strategies proposed here reduce the chances of transformer failure with the addition of plug-in electric vehicle loads, even for an under-designed transformer while uncontrolled and uncoordinated plug-in electric vehicle charging results in increased risk of transformer failure. - Highlights: • Charging algorithm for battery electric vehicles, for high penetration levels. • Algorithm reduces transformer overloading, for grid level valley filling. • Computation and communication requirements are minimal. • The distributed algorithm is implemented without large scale iterations. • Hot spot temperature and loss of life for transformers are evaluated.
International Nuclear Information System (INIS)
Jazebi, S.; Hosseinian, S.H.; Vahidi, B.
2011-01-01
Highlights: → Reconfiguration and DSTATCOM allocation are implemented for RDS planning. → Differential evolution algorithm is applied to solve the nonlinear problem. → Optimal status of tie switches, DSTATCOM size and location are determined. → The goal is to minimize network losses and to improve voltage profile. → The results show the effectiveness of the proposed method to satisfy objectives. -- Abstract: The main idea in distribution network reconfiguration is usually to reduce loss by changing the status of sectionalizing switches and determining appropriate tie switches. Recently Distribution FACTS (DFACTS) devices such as DSTATCOM also have been planned for loss reduction and voltage profile improvement in steady state conditions. This paper implements a combinatorial process based on reconfiguration and DSTATCOM allocation in order to mitigate losses and improve voltage profile in power distribution networks. The distribution system tie switches, DSTATCOM location and size have been optimally determined to obtain an appropriate operational condition. Differential evolution algorithm (DEA) has been used to solve and overcome the complicity of this combinatorial nonlinear optimization problem. To validate the accuracy of results a comparison with particle swarm optimization (PSO) has been made. Simulations have been applied on 69 and 83 busses distribution test systems. All optimization results show the effectiveness of the combinatorial approach in loss reduction and voltage profile improvement.
Vehicle Routing Problem Using Genetic Algorithm with Multi Compartment on Vegetable Distribution
Kurnia, Hari; Gustri Wahyuni, Elyza; Cergas Pembrani, Elang; Gardini, Syifa Tri; Kurnia Aditya, Silfa
2018-03-01
The problem that is often gained by the industries of managing and distributing vegetables is how to distribute vegetables so that the quality of the vegetables can be maintained properly. The problems encountered include optimal route selection and little travel time or so-called TSP (Traveling Salesman Problem). These problems can be modeled using the Vehicle Routing Problem (VRP) algorithm with rating ranking, a cross order based crossing, and also order based mutation mutations on selected chromosomes. This study uses limitations using only 20 market points, 2 point warehouse (multi compartment) and 5 vehicles. It is determined that for one distribution, one vehicle can only distribute to 4 market points only from 1 particular warehouse, and also one such vehicle can only accommodate 100 kg capacity.
Quantum Inequalities and Sequential Measurements
International Nuclear Information System (INIS)
Candelpergher, B.; Grandouz, T.; Rubinx, J.L.
2011-01-01
In this article, the peculiar context of sequential measurements is chosen in order to analyze the quantum specificity in the two most famous examples of Heisenberg and Bell inequalities: Results are found at some interesting variance with customary textbook materials, where the context of initial state re-initialization is described. A key-point of the analysis is the possibility of defining Joint Probability Distributions for sequential random variables associated to quantum operators. Within the sequential context, it is shown that Joint Probability Distributions can be defined in situations where not all of the quantum operators (corresponding to random variables) do commute two by two. (authors)
DEFF Research Database (Denmark)
Sun, Qiuye; Han, Renke; Zhang, Huaguang
2015-01-01
With the bidirectional power flow provided by the Energy Internet, various methods are promoted to improve and increase the energy utilization between Energy Internet and Main-Grid. This paper proposes a novel distributed coordinated controller combined with a multi-agent-based consensus algorithm...... which is applied to distributed generators in the Energy Internet. Then, the decomposed tasks, models, and information flow of the proposed method are analyzed. The proposed coordinated controller installed between the Energy Internet and the Main-Grid keeps voltage angles and amplitudes consensus while...... providing accurate power-sharing and minimizing circulating currents. Finally, the Energy Internet can be integrated into the Main-Grid seamlessly if necessary. Hence the Energy Internet can be operated as a spinning reserve system. Simulation results are provided to show the effectiveness of the proposed...
A Distributed and Energy-Efficient Algorithm for Event K-Coverage in Underwater Sensor Networks
Directory of Open Access Journals (Sweden)
Peng Jiang
2017-01-01
Full Text Available For event dynamic K-coverage algorithms, each management node selects its assistant node by using a greedy algorithm without considering the residual energy and situations in which a node is selected by several events. This approach affects network energy consumption and balance. Therefore, this study proposes a distributed and energy-efficient event K-coverage algorithm (DEEKA. After the network achieves 1-coverage, the nodes that detect the same event compete for the event management node with the number of candidate nodes and the average residual energy, as well as the distance to the event. Second, each management node estimates the probability of its neighbor nodes’ being selected by the event it manages with the distance level, the residual energy level, and the number of dynamic coverage event of these nodes. Third, each management node establishes an optimization model that uses expectation energy consumption and the residual energy variance of its neighbor nodes and detects the performance of the events it manages as targets. Finally, each management node uses a constrained non-dominated sorting genetic algorithm (NSGA-II to obtain the Pareto set of the model and the best strategy via technique for order preference by similarity to an ideal solution (TOPSIS. The algorithm first considers the effect of harsh underwater environments on information collection and transmission. It also considers the residual energy of a node and a situation in which the node is selected by several other events. Simulation results show that, unlike the on-demand variable sensing K-coverage algorithm, DEEKA balances and reduces network energy consumption, thereby prolonging the network’s best service quality and lifetime.
Capataz: a framework for distributing algorithms via the World Wide Web
Directory of Open Access Journals (Sweden)
Gonzalo J. Martínez
2015-08-01
Full Text Available In recent years, some scientists have embraced the distributed computing paradigm. As experiments and simulations demand ever more computing power, coordinating the efforts of many different processors is often the only reasonable resort. We developed an open-source distributed computing framework based on web technologies, and named it Capataz. Acting as an HTTP server, web browsers running on many different devices can connect to it to contribute in the execution of distributed algorithms written in Javascript. Capataz takes advantage of architectures with many cores using web workers. This paper presents an improvement in Capataz´ usability and why it was needed. In previous experiments the total time of distributed algorithms proved to be susceptible to changes in the execution time of the jobs. The system now adapts by bundling jobs together if they are too simple. The computational experiment to test the solution is a brute force estimation of pi. The benchmark results show that by bundling jobs, the overall perfomance is greatly increased.
Distributed Optimisation Algorithm for Demand Side Management in a Grid-Connected Smart Microgrid
Directory of Open Access Journals (Sweden)
Omowunmi Mary Longe
2017-06-01
Full Text Available The contributions of Distributed Energy Generation (DEG and Distributed Energy Storage (DES for Demand Side Management (DSM purposes in a smart macrogrid or microgrid cannot be over-emphasised. However, standalone DEG and DES can lead to under-utilisation of energy generation by consumers and financial investments; in grid-connection mode, though, DEG and DES can offer arbitrage opportunities for consumers and utility provider(s. A grid-connected smart microgrid comprising heterogeneous (active and passive smart consumers, electric vehicles and a large-scale centralised energy storage is considered in this paper. Efficient energy management by each smart entity is carried out by the proposed Microgrid Energy Management Distributed Optimisation Algorithm (MEM-DOA installed distributively within the network according to consumer type. Each smart consumer optimises its energy consumption and trading for comfort (demand satisfaction and profit. The proposed model was observed to yield better consumer satisfaction, higher financial savings, and reduced Peak-to-Average-Ratio (PAR demand on the utility grid. Other associated benefits of the model include reduced investment on peaker plants, grid reliability and environmental benefits. The MEM-DOA also offered participating smart consumers energy and tariff incentives so that passive smart consumers do not benefit more than active smart consumers, as was the case with some previous energy management algorithms.
Independent tasks scheduling in cloud computing via improved estimation of distribution algorithm
Sun, Haisheng; Xu, Rui; Chen, Huaping
2018-04-01
To minimize makespan for scheduling independent tasks in cloud computing, an improved estimation of distribution algorithm (IEDA) is proposed to tackle the investigated problem in this paper. Considering that the problem is concerned with multi-dimensional discrete problems, an improved population-based incremental learning (PBIL) algorithm is applied, which the parameter for each component is independent with other components in PBIL. In order to improve the performance of PBIL, on the one hand, the integer encoding scheme is used and the method of probability calculation of PBIL is improved by using the task average processing time; on the other hand, an effective adaptive learning rate function that related to the number of iterations is constructed to trade off the exploration and exploitation of IEDA. In addition, both enhanced Max-Min and Min-Min algorithms are properly introduced to form two initial individuals. In the proposed IEDA, an improved genetic algorithm (IGA) is applied to generate partial initial population by evolving two initial individuals and the rest of initial individuals are generated at random. Finally, the sampling process is divided into two parts including sampling by probabilistic model and IGA respectively. The experiment results show that the proposed IEDA not only gets better solution, but also has faster convergence speed.
Chatterjee, A.; Ghoshal, S. P.; Mukherjee, V.
In this paper, a conventional thermal power system equipped with automatic voltage regulator, IEEE type dual input power system stabilizer (PSS) PSS3B and integral controlled automatic generation control loop is considered. A distributed generation (DG) system consisting of aqua electrolyzer, photovoltaic cells, diesel engine generator, and some other energy storage devices like flywheel energy storage system and battery energy storage system is modeled. This hybrid distributed system is connected to the grid. While integrating this DG with the onventional thermal power system, improved transient performance is noticed. Further improvement in the transient performance of this grid connected DG is observed with the usage of superconducting magnetic energy storage device. The different tunable parameters of the proposed hybrid power system model are optimized by artificial bee colony (ABC) algorithm. The optimal solutions offered by the ABC algorithm are compared with those offered by genetic algorithm (GA). It is also revealed that the optimizing performance of the ABC is better than the GA for this specific application.
Directory of Open Access Journals (Sweden)
Abdallah A. Hassan
2014-12-01
Full Text Available Optimizing autonomous vehicle movements through roadway intersections is a challenging problem. It has been demonstrated in the literature that traditional traffic control, such as traffic signal and stop sign control are not optimal especially for heavy traffic demand levels. Alternatively, centralized autonomous vehicle control strategies are costly and not scalable given that the ability of a central controller to track and schedule the movement of hundreds of vehicles in real-time is questionable. Consequently, in this paper a fully distributed algorithm is proposed where vehicles in the vicinity of an intersection continuously cooperate with each other to develop a schedule that allows them to safely proceed through the intersection while incurring minimum delay. Unlike other distributed approaches described in the literature, the wireless communication constraints are considered in the design of the control algorithm. Specifically, the proposed algorithm requires vehicles heading to an intersection to communicate only with neighboring vehicles, while the lead vehicles on each approach lane share information to develop a complete intersection utilization schedule. The scheduling rotates between vehicles to identify higher traffic volumes and favor vehicles coming from heavier lanes to minimize the overall intersection delay. The simulated experiments show significant reductions in the average delay using the proposed approach compared to other methods reported in the literature and reduction in the maximum delay experienced by a vehicle especially in cases of heavy traffic demand levels.
International Nuclear Information System (INIS)
Moretto, L.G.
1977-08-01
The hierarchy of the collective relaxation times in heavy ion reactions is briefly reviewed. An improved diffusion model is introduced and applied to interpret the fragment Z and angular distributions for some typical reactions. The equilibrium in the neutron-to-proton ratio as well as the sharing of the excitation energy between fragments is studied by a coincidence method which leads to the measurement of the charge, mass and mean number of nucleons emitted by each fragment. The final destiny of the dissipative energy is determined by measuring the atomic number of two coincident fragments, thus obtaining the missing charge as a function of bombarding energy and the Q of the reaction. The sequential fission probability of the heavy recoil is established as a function of the Z and kinetic energy of the light partner. The out-of-plane angular distribution of the fission fragments is correlated with the fissionability and interpreted in terms of various sources of angular momentum misalignment. The γ-ray multiplicities and the γ-ray angular distributions associated with deep inelastic event are discussed in terms of the angular momentum transfer and in terms of the diffusion model
Energy Technology Data Exchange (ETDEWEB)
Loring, Burlen; Karimabadi, Homa; Rortershteyn, Vadim
2014-07-01
The surface line integral convolution(LIC) visualization technique produces dense visualization of vector fields on arbitrary surfaces. We present a screen space surface LIC algorithm for use in distributed memory data parallel sort last rendering infrastructures. The motivations for our work are to support analysis of datasets that are too large to fit in the main memory of a single computer and compatibility with prevalent parallel scientific visualization tools such as ParaView and VisIt. By working in screen space using OpenGL we can leverage the computational power of GPUs when they are available and run without them when they are not. We address efficiency and performance issues that arise from the transformation of data from physical to screen space by selecting an alternate screen space domain decomposition. We analyze the algorithm's scaling behavior with and without GPUs on two high performance computing systems using data from turbulent plasma simulations.
Distributed optical fiber vibration sensing using phase-generated carrier demodulation algorithm
Yu, Zhihua; Zhang, Qi; Zhang, Mingyu; Dai, Haolong; Zhang, Jingjing; Liu, Li; Zhang, Lijun; Jin, Xing; Wang, Gaifang; Qi, Guang
2018-05-01
A novel optical fiber-distributed vibration-sensing system is proposed, which is based on self-interference of Rayleigh backscattering with phase-generated carrier (PGC) demodulation algorithm. Pulsed lights are sent into the sensing fiber and the Rayleigh backscattering light from a certain position along the sensing fiber would interfere through an unbalanced Michelson interferometry to generate the interference light. An improved PGC demodulation algorithm is carried out to recover the phase information of the interference signal, which carries the sensing information. Three vibration events were applied simultaneously to different positions over 2000 m sensing fiber and demodulated correctly. The spatial resolution is 10 m, and the noise level of the Φ-OTDR system we proposed is about 10-3 rad/\\surd {Hz}, and the signal-to-noise ratio is about 30.34 dB.
Peterson, Harold; Koshak, William J.
2009-01-01
An algorithm has been developed to estimate the altitude distribution of one-meter lightning channel segments. The algorithm is required as part of a broader objective that involves improving the lightning NOx emission inventories of both regional air quality and global chemistry/climate models. The algorithm was tested and applied to VHF signals detected by the North Alabama Lightning Mapping Array (NALMA). The accuracy of the algorithm was characterized by comparing algorithm output to the plots of individual discharges whose lengths were computed by hand; VHF source amplitude thresholding and smoothing were applied to optimize results. Several thousands of lightning flashes within 120 km of the NALMA network centroid were gathered from all four seasons, and were analyzed by the algorithm. The mean, standard deviation, and median statistics were obtained for all the flashes, the ground flashes, and the cloud flashes. One-meter channel segment altitude distributions were also obtained for the different seasons.
Random sequential adsorption of cubes
Cieśla, Michał; Kubala, Piotr
2018-01-01
Random packings built of cubes are studied numerically using a random sequential adsorption algorithm. To compare the obtained results with previous reports, three different models of cube orientation sampling were used. Also, three different cube-cube intersection algorithms were tested to find the most efficient one. The study focuses on the mean saturated packing fraction as well as kinetics of packing growth. Microstructural properties of packings were analyzed using density autocorrelation function.
A partition enhanced mining algorithm for distributed association rule mining systems
Directory of Open Access Journals (Sweden)
A.O. Ogunde
2015-11-01
Full Text Available The extraction of patterns and rules from large distributed databases through existing Distributed Association Rule Mining (DARM systems is still faced with enormous challenges such as high response times, high communication costs and inability to adapt to the constantly changing databases. In this work, a Partition Enhanced Mining Algorithm (PEMA is presented to address these problems. In PEMA, the Association Rule Mining Coordinating Agent receives a request and decides the appropriate data sites, partitioning strategy and mining agents to use. The mining process is divided into two stages. In the first stage, the data agents horizontally segment the databases with small average transaction length into relatively smaller partitions based on the number of available sites and the available memory. On the other hand, databases with relatively large average transaction length were vertically partitioned. After this, Mobile Agent-Based Association Rule Mining-Agents, which are the mining agents, carry out the discovery of the local frequent itemsets. At the second stage, the local frequent itemsets were incrementally integrated by the from one data site to another to get the global frequent itemsets. This reduced the response time and communication cost in the system. Results from experiments conducted on real datasets showed that the average response time of PEMA showed an improvement over existing algorithms. Similarly, PEMA incurred lower communication costs with average size of messages exchanged lower when compared with benchmark DARM systems. This result showed that PEMA could be efficiently deployed for efficient discovery of valuable knowledge in distributed databases.
Energy Technology Data Exchange (ETDEWEB)
Dall-Anese, Emiliano [National Renewable Energy Laboratory (NREL), Golden, CO (United States); Zhou, Xinyang [University of Colorado; Liu, Zhiyuan [University of Colorado; Chen, Lijun [University of Colorado
2017-10-03
This paper considers distribution networks with distributed energy resources and discrete-rate loads, and designs an incentive-based algorithm that allows the network operator and the customers to pursue given operational and economic objectives, while concurrently ensuring that voltages are within prescribed limits. Four major challenges include: (1) the non-convexity from discrete decision variables, (2) the non-convexity due to a Stackelberg game structure, (3) unavailable private information from customers, and (4) different update frequency from two types of devices. In this paper, we first make convex relaxation for discrete variables, then reformulate the non-convex structure into a convex optimization problem together with pricing/reward signal design, and propose a distributed stochastic dual algorithm for solving the reformulated problem while restoring feasible power rates for discrete devices. By doing so, we are able to statistically achieve the solution of the reformulated problem without exposure of any private information from customers. Stability of the proposed schemes is analytically established and numerically corroborated.
He, Zhenzong; Qi, Hong; Yao, Yuchen; Ruan, Liming
2014-12-01
The Ant Colony Optimization algorithm based on the probability density function (PDF-ACO) is applied to estimate the bimodal aerosol particle size distribution (PSD). The direct problem is solved by the modified Anomalous Diffraction Approximation (ADA, as an approximation for optically large and soft spheres, i.e., χ⪢1 and |m-1|⪡1) and the Beer-Lambert law. First, a popular bimodal aerosol PSD and three other bimodal PSDs are retrieved in the dependent model by the multi-wavelength extinction technique. All the results reveal that the PDF-ACO algorithm can be used as an effective technique to investigate the bimodal PSD. Then, the Johnson's SB (J-SB) function and the modified beta (M-β) function are employed as the general distribution function to retrieve the bimodal PSDs under the independent model. Finally, the J-SB and M-β functions are applied to recover actual measurement aerosol PSDs over Beijing and Shanghai obtained from the aerosol robotic network (AERONET). The numerical simulation and experimental results demonstrate that these two general functions, especially the J-SB function, can be used as a versatile distribution function to retrieve the bimodal aerosol PSD when no priori information about the PSD is available.
Directory of Open Access Journals (Sweden)
Álvaro Gutiérrez
2011-11-01
Full Text Available Swarms of robots can use their sensing abilities to explore unknown environments and deploy on sites of interest. In this task, a large number of robots is more effective than a single unit because of their ability to quickly cover the area. However, the coordination of large teams of robots is not an easy problem, especially when the resources for the deployment are limited. In this paper, the Distributed Bees Algorithm (DBA, previously proposed by the authors, is optimized and applied to distributed target allocation in swarms of robots. Improved target allocation in terms of deployment cost efficiency is achieved through optimization of the DBA’s control parameters by means of a Genetic Algorithm. Experimental results show that with the optimized set of parameters, the deployment cost measured as the average distance traveled by the robots is reduced. The cost-efficient deployment is in some cases achieved at the expense of increased robots’ distribution error. Nevertheless, the proposed approach allows the swarm to adapt to the operating conditions when available resources are scarce.
A Power Load Distribution Algorithm to Optimize Data Center Electrical Flow
Directory of Open Access Journals (Sweden)
Paulo Maciel
2013-07-01
Full Text Available Energy consumption is a matter of common concern in the world today. Research demonstrates that as a consequence of the constantly evolving and expanding field of information technology, data centers are now major consumers of electrical energy. Such high electrical energy consumption emphasizes the issues of sustainability and cost. Against this background, the present paper proposes a power load distribution algorithm (PLDA to optimize energy distribution of data center power infrastructures. The PLDA, which is based on the Ford-Fulkerson algorithm, is supported by an environment called ASTRO, capable of performing the integrated evaluation of dependability, cost and sustainability. More specifically, the PLDA optimizes the flow distribution of the energy flow model (EFM. EFMs are responsible for estimating sustainability and cost issues of data center infrastructures without crossing the restrictions of the power capacity that each device can provide (power system or extract (cooling system. Additionally, a case study is presented that analyzed seven data center power architectures. Significant results were observed, achieving a reduction in power consumption of up to 15.5%.
Simulation based sequential Monte Carlo methods for discretely observed Markov processes
Neal, Peter
2014-01-01
Parameter estimation for discretely observed Markov processes is a challenging problem. However, simulation of Markov processes is straightforward using the Gillespie algorithm. We exploit this ease of simulation to develop an effective sequential Monte Carlo (SMC) algorithm for obtaining samples from the posterior distribution of the parameters. In particular, we introduce two key innovations, coupled simulations, which allow us to study multiple parameter values on the basis of a single sim...
Sequential Change-Point Detection via Online Convex Optimization
Directory of Open Access Journals (Sweden)
Yang Cao
2018-02-01
Full Text Available Sequential change-point detection when the distribution parameters are unknown is a fundamental problem in statistics and machine learning. When the post-change parameters are unknown, we consider a set of detection procedures based on sequential likelihood ratios with non-anticipating estimators constructed using online convex optimization algorithms such as online mirror descent, which provides a more versatile approach to tackling complex situations where recursive maximum likelihood estimators cannot be found. When the underlying distributions belong to a exponential family and the estimators satisfy the logarithm regret property, we show that this approach is nearly second-order asymptotically optimal. This means that the upper bound for the false alarm rate of the algorithm (measured by the average-run-length meets the lower bound asymptotically up to a log-log factor when the threshold tends to infinity. Our proof is achieved by making a connection between sequential change-point and online convex optimization and leveraging the logarithmic regret bound property of online mirror descent algorithm. Numerical and real data examples validate our theory.
International Nuclear Information System (INIS)
Hung, Shih-Yu; Shen, Ming-Ho; Chang, Ying-Pin
2009-01-01
The sequential neural-network approximation and orthogonal array (SNAOA) were used to shorten the cooling time for the rapid cooling process such that the normalized maximum resolved stress in silicon wafer was always below one in this study. An orthogonal array was first conducted to obtain the initial solution set. The initial solution set was treated as the initial training sample. Next, a back-propagation sequential neural network was trained to simulate the feasible domain to obtain the optimal parameter setting. The size of the training sample was greatly reduced due to the use of the orthogonal array. In addition, a restart strategy was also incorporated into the SNAOA so that the searching process may have a better opportunity to reach a near global optimum. In this work, we considered three different cooling control schemes during the rapid thermal process: (1) downward axial gas flow cooling scheme; (2) upward axial gas flow cooling scheme; (3) dual axial gas flow cooling scheme. Based on the maximum shear stress failure criterion, the other control factors such as flow rate, inlet diameter, outlet width, chamber height and chamber diameter were also examined with respect to cooling time. The results showed that the cooling time could be significantly reduced using the SNAOA approach
Day-ahead distributed energy resource scheduling using differential search algorithm
DEFF Research Database (Denmark)
Soares, J.; Lobo, C.; Silva, M.
2015-01-01
The number of dispersed energy resources is growing every day, such as the use of more distributed generators. This paper deals with energy resource scheduling model in future smart grids. The methodology can be used by virtual power players (VPPs) considering day-ahead time horizon. This method...... considers that energy resources are managed by a VPP which establishes contracts with their owners. The full AC power flow calculation included in the model takes into account network constraints. This paper presents an application of differential search algorithm (DSA) for solving the day-ahead scheduling...
Distribution Network Expansion Planning Based on Multi-objective PSO Algorithm
DEFF Research Database (Denmark)
Zhang, Chunyu; Ding, Yi; Wu, Qiuwei
2013-01-01
This paper presents a novel approach for electrical distribution network expansion planning using multi-objective particle swarm optimization (PSO). The optimization objectives are: investment and operation cost, energy losses cost, and power congestion cost. A two-phase multi-objective PSO...... algorithm was proposed to solve this optimization problem, which can accelerate the convergence and guarantee the diversity of Pareto-optimal front set as well. The feasibility and effectiveness of both the proposed multi-objective planning approach and the improved multi-objective PSO have been verified...
Inverse estimation of the particle size distribution using the Fruit Fly Optimization Algorithm
International Nuclear Information System (INIS)
He, Zhenzong; Qi, Hong; Yao, Yuchen; Ruan, Liming
2015-01-01
The Fruit Fly Optimization Algorithm (FOA) is applied to retrieve the particle size distribution (PSD) for the first time. The direct problems are solved by the modified Anomalous Diffraction Approximation (ADA) and the Lambert–Beer Law. Firstly, three commonly used monomodal PSDs, i.e. the Rosin–Rammer (R–R) distribution, the normal (N–N) distribution and the logarithmic normal (L–N) distribution, and the bimodal Rosin–Rammer distribution function are estimated in the dependent model. All the results show that the FOA can be used as an effective technique to estimate the PSDs under the dependent model. Then, an optimal wavelength selection technique is proposed to improve the retrieval results of bimodal PSD. Finally, combined with two general functions, i.e. the Johnson's S B (J-S B ) function and the modified beta (M-β) function, the FOA is employed to recover actual measurement aerosol PSDs over Beijing and Hangzhou obtained from the aerosol robotic network (AERONET). All the numerical simulations and experiment results demonstrate that the FOA can be used to retrieve actual measurement PSDs, and more reliable and accurate results can be obtained, if the J-S B function is employed
Energy Technology Data Exchange (ETDEWEB)
Cebrian, Juan Carlos; Kagan, Nelson [Department of Electrical Engineering, University of Sao Paulo, Escola Politecnica, Av. Prof. Luciano Gualberto, travessa 3 n 380 - CEP - 05508-970 - Sao Paulo (Brazil)
2010-01-15
In this paper a computational implementation of an evolutionary algorithm (EA) is shown in order to tackle the problem of reconfiguring radial distribution systems. The developed module considers power quality indices such as long duration interruptions and customer process disruptions due to voltage sags, by using the Monte Carlo simulation method. Power quality costs are modeled into the mathematical problem formulation, which are added to the cost of network losses. As for the EA codification proposed, a decimal representation is used. The EA operators, namely selection, recombination and mutation, which are considered for the reconfiguration algorithm, are herein analyzed. A number of selection procedures are analyzed, namely tournament, elitism and a mixed technique using both elitism and tournament. The recombination operator was developed by considering a chromosome structure representation that maps the network branches and system radiality, and another structure that takes into account the network topology and feasibility of network operation to exchange genetic material. The topologies regarding the initial population are randomly produced so as radial configurations are produced through the Prim and Kruskal algorithms that rapidly build minimum spanning trees. (author)
Patent Keyword Extraction Algorithm Based on Distributed Representation for Patent Classification
Directory of Open Access Journals (Sweden)
Jie Hu
2018-02-01
Full Text Available Many text mining tasks such as text retrieval, text summarization, and text comparisons depend on the extraction of representative keywords from the main text. Most existing keyword extraction algorithms are based on discrete bag-of-words type of word representation of the text. In this paper, we propose a patent keyword extraction algorithm (PKEA based on the distributed Skip-gram model for patent classification. We also develop a set of quantitative performance measures for keyword extraction evaluation based on information gain and cross-validation, based on Support Vector Machine (SVM classification, which are valuable when human-annotated keywords are not available. We used a standard benchmark dataset and a homemade patent dataset to evaluate the performance of PKEA. Our patent dataset includes 2500 patents from five distinct technological fields related to autonomous cars (GPS systems, lidar systems, object recognition systems, radar systems, and vehicle control systems. We compared our method with Frequency, Term Frequency-Inverse Document Frequency (TF-IDF, TextRank and Rapid Automatic Keyword Extraction (RAKE. The experimental results show that our proposed algorithm provides a promising way to extract keywords from patent texts for patent classification.
A Unified Algorithm for Virtual Desktops Placement in Distributed Cloud Computing
Directory of Open Access Journals (Sweden)
Jiangtao Zhang
2016-01-01
Full Text Available Distributed cloud has been widely adopted to support service requests from dispersed regions, especially for large enterprise which requests virtual desktops for multiple geodistributed branch companies. The cloud service provider (CSP aims to deliver satisfactory services at the least cost. CSP selects proper data centers (DCs closer to the branch companies so as to shorten the response time to user request. At the same time, it also strives to cut cost considering both DC level and server level. At DC level, the expensive long distance inter-DC bandwidth consumption should be reduced and lower electricity price is sought. Inside each tree-like DC, servers are trying to be used as little as possible so as to save equipment cost and power. In nature, there is a noncooperative relation between the DC level and server level in the selection. To attain these objectives and capture the noncooperative relation, multiobjective bilevel programming is used to formulate the problem. Then a unified genetic algorithm is proposed to solve the problem which realizes the selection of DC and server simultaneously. The extensive simulation shows that the proposed algorithm outperforms baseline algorithm in both quality of service guaranteeing and cost saving.
Directory of Open Access Journals (Sweden)
Han Liwei
2014-07-01
Full Text Available Monitoring data on an earth-rockfill dam constitutes a form of spatial data. Such data include much uncertainty owing to the limitation of measurement information, material parameters, load, geometry size, initial conditions, boundary conditions and the calculation model. So the cloud probability density of the monitoring data must be addressed. In this paper, the cloud theory model was used to address the uncertainty transition between the qualitative concept and the quantitative description. Then an improved algorithm of cloud probability distribution density based on a backward cloud generator was proposed. This was used to effectively convert certain parcels of accurate data into concepts which can be described by proper qualitative linguistic values. Such qualitative description was addressed as cloud numerical characteristics-- {Ex, En, He}, which could represent the characteristics of all cloud drops. The algorithm was then applied to analyze the observation data of a piezometric tube in an earth-rockfill dam. And experiment results proved that the proposed algorithm was feasible, through which, we could reveal the changing regularity of piezometric tube’s water level. And the damage of the seepage in the body was able to be found out.
Algorithm of Data Reduce in Determination of Aerosol Particle Size Distribution at Damps/C
International Nuclear Information System (INIS)
Muhammad-Priyatna; Otto-Pribadi-Ruslanto
2001-01-01
The analysis had to do for algorithm of data reduction on Damps/C (Differential Mobility Particle Sizer with Condensation Particle Counter) system, this is for determine aerosol particle size distribution with range 0,01 μm to 1 μm in diameter. Damps/C (Differential Mobility Particle Sizer with Condensation Particle Counter) system contents are software and hardware. The hardware used determine of mobilities of aerosol particle and so the software used determine aerosol particle size distribution in diameter. The mobilities and diameter particle had connection in the electricity field. That is basic program for reduction of data and particle size conversion from particle mobility become particle diameter. The analysis to get transfer function value, Ω, is 0.5. The data reduction program to do conversation mobility basis become diameter basis with number efficiency correction, transfer function value, and poly charge particle. (author)
Energy Technology Data Exchange (ETDEWEB)
Schatz, Martin D. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); Kolda, Tamara G. [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States); van de Geijn, Robert [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)
2015-09-01
Large-scale datasets in computational chemistry typically require distributed-memory parallel methods to perform a special operation known as tensor contraction. Tensors are multidimensional arrays, and a tensor contraction is akin to matrix multiplication with special types of permutations. Creating an efficient algorithm and optimized im- plementation in this domain is complex, tedious, and error-prone. To address this, we develop a notation to express data distributions so that we can apply use automated methods to find optimized implementations for tensor contractions. We consider the spin-adapted coupled cluster singles and doubles method from computational chemistry and use our methodology to produce an efficient implementation. Experiments per- formed on the IBM Blue Gene/Q and Cray XC30 demonstrate impact both improved performance and reduced memory consumption.
Reconfiguration of Smart Distribution Network in the Presence of Renewable DG’s Using GWO Algorithm
Siavash, M.; Pfeifer, C.; Rahiminejad, A.; Vahidi, B.
2017-08-01
In this paper, the optimal reconfiguration of smart distribution system is performed with the aim of active power loss reduction and voltage stability improvement. The distribution network is considered equipped with wind turbines and solar cells as Renewable DG’s (RDG’s). Because of the presence of smart metering devices, the network state is known accurately at any moment. Based on the network conditions (the amount of load and generation of RDG’s), the optimal configuration of the network is obtained. The optimization problem is solved using a recently introduced method known as Grey Wolf Optimizer (GWO). The proposed approach is applied on 69-bus radial test system and the results of the GWO are compared to those of Particle Swarm Optimization (PSO), and Genetic Algorithm (GA). The results show the effectiveness of the proposed approach and the selected optimization method.
A new multiple robot path planning algorithm: dynamic distributed particle swarm optimization.
Ayari, Asma; Bouamama, Sadok
2017-01-01
Multiple robot systems have become a major study concern in the field of robotic research. Their control becomes unreliable and even infeasible if the number of robots increases. In this paper, a new dynamic distributed particle swarm optimization (D 2 PSO) algorithm is proposed for trajectory path planning of multiple robots in order to find collision-free optimal path for each robot in the environment. The proposed approach consists in calculating two local optima detectors, LOD pBest and LOD gBest . Particles which are unable to improve their personal best and global best for predefined number of successive iterations would be replaced with restructured ones. Stagnation and local optima problems would be avoided by adding diversity to the population, without losing the fast convergence characteristic of PSO. Experiments with multiple robots are provided and proved effectiveness of such approach compared with the distributed PSO.
Directory of Open Access Journals (Sweden)
Mi Jeong Kim
2015-07-01
Full Text Available In this paper, a distributed synchronization technique based on a bio-inspired algorithm is proposed for an orthogonal frequency division multiple access (OFDMA-based wireless mesh network (WMN with a time difference of arrival. The proposed time- and frequency-synchronization technique uses only the signals received from the neighbor nodes, by considering the effect of the propagation delay between the nodes. It achieves a fast synchronization with a relatively low computational complexity because it is operated in a distributed manner, not requiring any feedback channel for the compensation of the propagation delays. In addition, a self-organization scheme that can be effectively used to construct 1-hop neighbor nodes is proposed for an OFDMA-based WMN with a large number of nodes. The performance of the proposed technique is evaluated with regard to the convergence property and synchronization success probability using a computer simulation.
Kim, Mi Jeong; Maeng, Sung Joon; Cho, Yong Soo
2015-07-28
In this paper, a distributed synchronization technique based on a bio-inspired algorithm is proposed for an orthogonal frequency division multiple access (OFDMA)-based wireless mesh network (WMN) with a time difference of arrival. The proposed time- and frequency-synchronization technique uses only the signals received from the neighbor nodes, by considering the effect of the propagation delay between the nodes. It achieves a fast synchronization with a relatively low computational complexity because it is operated in a distributed manner, not requiring any feedback channel for the compensation of the propagation delays. In addition, a self-organization scheme that can be effectively used to construct 1-hop neighbor nodes is proposed for an OFDMA-based WMN with a large number of nodes. The performance of the proposed technique is evaluated with regard to the convergence property and synchronization success probability using a computer simulation.
International Nuclear Information System (INIS)
Chan, Apple L.S.; Hanby, Vic I.; Chow, T.T.
2007-01-01
A district cooling system is a sustainable means of distribution of cooling energy through mass production. A cooling medium like chilled water is generated at a central refrigeration plant and supplied to serve a group of consumer buildings through a piping network. Because of the substantial capital investment involved, an optimal design of the distribution piping configuration is one of the crucial factors for successful implementation of the district cooling scheme. In the present study, genetic algorithm (GA) incorporated with local search techniques was developed to find the optimal/near optimal configuration of the piping network in a hypothetical site. The effect of local search, mutation rate and frequency of local search on the performance of the GA in terms of both solution quality and computation time were investigated and presented in this paper
A Matrix-Based Proactive Data Relay Algorithm for Large Distributed Sensor Networks.
Xu, Yang; Hu, Xuemei; Hu, Haixiao; Liu, Ming
2016-08-16
In large-scale distributed sensor networks, sensed data is required to be relayed around the network so that one or few sensors can gather adequate relative data to produce high quality information for decision-making. In regards to very high energy-constraint sensor nodes, data transmission should be extremely economical. However, traditional data delivery protocols are potentially inefficient relaying unpredictable sensor readings for data fusion in large distributed networks for either overwhelming query transmissions or unnecessary data coverage. By building sensors' local model from their previously transmitted data in three matrixes, we have developed a novel energy-saving data relay algorithm, which allows sensors to proactively make broadcast decisions by using a neat matrix computation to provide balance between transmission and energy-saving. In addition, we designed a heuristic maintenance algorithm to efficiently update these three matrices. This can easily be deployed to large-scale mobile networks in which decisions of sensors are based on their local matrix models no matter how large the network is, and the local models of these sensors are updated constantly. Compared with some traditional approaches based on our simulations, the efficiency of this approach is manifested in uncertain environment. The results show that our approach is scalable and can effectively balance aggregating data with minimizing energy consumption.
A Matrix-Based Proactive Data Relay Algorithm for Large Distributed Sensor Networks
Directory of Open Access Journals (Sweden)
Yang Xu
2016-08-01
Full Text Available In large-scale distributed sensor networks, sensed data is required to be relayed around the network so that one or few sensors can gather adequate relative data to produce high quality information for decision-making. In regards to very high energy-constraint sensor nodes, data transmission should be extremely economical. However, traditional data delivery protocols are potentially inefficient relaying unpredictable sensor readings for data fusion in large distributed networks for either overwhelming query transmissions or unnecessary data coverage. By building sensors’ local model from their previously transmitted data in three matrixes, we have developed a novel energy-saving data relay algorithm, which allows sensors to proactively make broadcast decisions by using a neat matrix computation to provide balance between transmission and energy-saving. In addition, we designed a heuristic maintenance algorithm to efficiently update these three matrices. This can easily be deployed to large-scale mobile networks in which decisions of sensors are based on their local matrix models no matter how large the network is, and the local models of these sensors are updated constantly. Compared with some traditional approaches based on our simulations, the efficiency of this approach is manifested in uncertain environment. The results show that our approach is scalable and can effectively balance aggregating data with minimizing energy consumption.
Algorithm and Implementation of Distributed ESN Using Spark Framework and Parallel PSO
Directory of Open Access Journals (Sweden)
Kehe Wu
2017-04-01
Full Text Available The echo state network (ESN employs a huge reservoir with sparsely and randomly connected internal nodes and only trains the output weights, which avoids the suboptimal problem, exploding and vanishing gradients, high complexity and other disadvantages faced by traditional recurrent neural network (RNN training. In light of the outstanding adaption to nonlinear dynamical systems, ESN has been applied into a wide range of applications. However, in the era of Big Data, with an enormous amount of data being generated continuously every day, the data are often distributed and stored in real applications, and thus the centralized ESN training process is prone to being technologically unsuitable. In order to achieve the requirement of Big Data applications in the real world, in this study we propose an algorithm and its implementation for distributed ESN training. The mentioned algorithm is based on the parallel particle swarm optimization (P-PSO technique and the implementation uses Spark, a famous large-scale data processing framework. Four extremely large-scale datasets, including artificial benchmarks, real-world data and image data, are adopted to verify our framework on a stretchable platform. Experimental results indicate that the proposed work is accurate in the era of Big Data, regarding speed, accuracy and generalization capabilities.
Baldini, Luca; Adirosi, Elisa; Roberto, Nicoletta; Vulpiani, Gianfranco; Russo, Fabio; Napolitano, Francesco
2015-04-01
Radar precipitation retrieval uses several relationships that parameterize precipitation properties (like rainfall rate and liquid water content and attenuation (in case of radars at attenuated frequencies such as those at C- and X- band) as a function of combinations of radar measurements. The uncertainty in such relations highly affects the uncertainty precipitation and attenuation estimates. A commonly used method to derive such relationships is to apply regression methods to precipitation measurements and radar observables simulated from datasets of drop size distributions (DSD) using microphysical and electromagnetic assumptions. DSD datasets are determined both by theoretical considerations (i.e. based on the assumption that the radar always samples raindrops whose sizes follow a gamma distribution) or from experimental measurements collected throughout the years by disdrometers. In principle, using long-term disdrometer measurements provide parameterizations more representative of a specific climatology. However, instrumental errors, specific of a disdrometer, can affect the results. In this study, different weather radar algorithms resulting from DSDs collected by diverse types of disdrometers, namely 2D video disdrometer, first and second generation of OTT Parsivel laser disdrometer, and Thies Clima laser disdrometer, in the area of Rome (Italy) are presented and discussed to establish at what extent dual-polarization radar algorithms derived from experimental DSD datasets are influenced by the different error structure of the different type of disdrometers used to collect the data.
Directory of Open Access Journals (Sweden)
Daniela Saturnino
2008-10-01
Full Text Available The issue of adaptive and distributed cross-layer resource allocation for energy efficiency in uplink code-division multiple-access (CDMA wireless data networks is addressed. The resource allocation problems are formulated as noncooperative games wherein each terminal seeks to maximize its own energy efficiency, namely, the number of reliably transmitted information symbols per unit of energy used for transmission. The focus of this paper is on the issue of adaptive and distributed implementation of policies arising from this approach, that is, it is assumed that only readily available measurements, such as the received data, are available at the receiver in order to play the considered games. Both single-cell and multicell networks are considered. Stochastic implementations of noncooperative games for power allocation, spreading code allocation, and choice of the uplink (linear receiver are thus proposed, and analytical results describing the convergence properties of selected stochastic algorithms are also given. Extensive simulation results show that, in many instances of practical interest, the proposed stochastic algorithms approach with satisfactory accuracy the performance of nonadaptive games, whose implementation requires much more prior information.
Directory of Open Access Journals (Sweden)
Jinhyun Park
2015-08-01
Full Text Available The in-wheel electric vehicle is expected to be a popular next-generation vehicle because an in-wheel system can simplify the powertrain and improve driving performance. In addition, it also has an advantage in that it maximizes driving efficiency through independent torque control considering the motor efficiency. However, there is an instability problem if only the driving torque is controlled in consideration of only the motor efficiency. In this paper, integrated torque distribution strategies are proposed to overcome these problems. The control algorithm consists of various strategies for optimizing driving efficiency, satisfying driver demands, and considering tire slip and vehicle cornering. Fuzzy logic is used to determine the appropriate timing of intervention for each distribution strategy. A performance simulator for in-wheel electric vehicles was developed by using MATLAB/Simulink and CarSim to validate the control strategies. From simulation results under complex driving conditions, the proposed algorithm was verified to improve both the driving stability and fuel economy of the in-wheel vehicle.
Directory of Open Access Journals (Sweden)
Litian Duan
2016-11-01
Full Text Available In the multiple-reader environment (MRE of radio frequency identification (RFID system, multiple readers are often scheduled to interrogate the randomized tags via operating at different time slots or frequency channels to decrease the signal interferences. Based on this, a Geometric Distribution-based Multiple-reader Scheduling Optimization Algorithm using Artificial Immune System (GD-MRSOA-AIS is proposed to fairly and optimally schedule the readers operating from the viewpoint of resource allocations. GD-MRSOA-AIS is composed of two parts, where a geometric distribution function combined with the fairness consideration is first introduced to generate the feasible scheduling schemes for reader operation. After that, artificial immune system (including immune clone, immune mutation and immune suppression quickly optimize these feasible ones as the optimal scheduling scheme to ensure that readers are fairly operating with larger effective interrogation range and lower interferences. Compared with the state-of-the-art algorithm, the simulation results indicate that GD-MRSOA-AIS could efficiently schedules the multiple readers operating with a fairer resource allocation scheme, performing in larger effective interrogation range.
TRMM Satellite Algorithm Estimates to Represent the Spatial Distribution of Rainstorms
Directory of Open Access Journals (Sweden)
Patrick Marina
2017-01-01
Full Text Available On-site measurements from rain gauge provide important information for the design, construction, and operation of water resources engineering projects, groundwater potentials, and the water supply and irrigation systems. A dense gauging network is needed to accurately characterize the variation of rainfall over a region, unfitting for conditions with limited networks, such as in Sarawak, Malaysia. Hence, satellite-based algorithm estimates are introduced as an innovative solution to these challenges. With accessibility to dataset retrievals from public domain websites, it has become a useful source to measure rainfall for a wider coverage area at finer temporal resolution. This paper aims to investigate the rainfall estimates prepared by Tropical Rainfall Measuring Mission (TRMM to explain whether it is suitable to represent the distribution of extreme rainfall in Sungai Sarawak Basin. Based on the findings, more uniform correlations for the investigated storms can be observed for low to medium altitude (>40 MASL. It is found for the investigated events of Jan 05-11, 2009: the normalized root mean square error (NRMSE = 36.7 %; and good correlation (CC = 0.9. These findings suggest that satellite algorithm estimations from TRMM are suitable to represent the spatial distribution of extreme rainfall.
Shao, Zhongshi; Pi, Dechang; Shao, Weishi
2018-05-01
This article presents an effective estimation of distribution algorithm, named P-EDA, to solve the blocking flow-shop scheduling problem (BFSP) with the makespan criterion. In the P-EDA, a Nawaz-Enscore-Ham (NEH)-based heuristic and the random method are combined to generate the initial population. Based on several superior individuals provided by a modified linear rank selection, a probabilistic model is constructed to describe the probabilistic distribution of the promising solution space. The path relinking technique is incorporated into EDA to avoid blindness of the search and improve the convergence property. A modified referenced local search is designed to enhance the local exploitation. Moreover, a diversity-maintaining scheme is introduced into EDA to avoid deterioration of the population. Finally, the parameters of the proposed P-EDA are calibrated using a design of experiments approach. Simulation results and comparisons with some well-performing algorithms demonstrate the effectiveness of the P-EDA for solving BFSP.
Directory of Open Access Journals (Sweden)
UMADEVI K. S.
2017-07-01
Full Text Available WiMedia Medium Access Control (MAC provides high rate data transfer for wireless networking thereby enables construction of high speed home networks. It facilitates data communication between the nodes through two modes namely: i Distributed Reservation Protocol (DRP for isochronous traffic and ii Prioritized Contention Access (PCA for asynchronous traffic. PCA mode enables medium access using CSMA/CA similar to IEEE 802.11e. In the presence of DRP, the throughput of PCA saturates when there is an increase in the number of devices accessing PCA channel. Researchers suggest that the better utilization of medium resolves many issues in an effective way. To demonstrate the effective utilization of the medium, Contention Based Distributed Reservation Protocol Allocation Algorithm for WiMedia Medium Access Control is proposed for reserving Medium Access Slots under DRP in the presence of PCA. The proposed algorithm provides a better medium access, reduces energy consumption and enhances the throughput when compared to the existing methodologies.
Study on two-dimensional distribution of X-ray image based on improved Elman algorithm
International Nuclear Information System (INIS)
Wang, Fang; Wang, Ming-Yuan; Tian, Feng-Shuo; Liu, Yu-Fang; Li, Lei; Zhao, Jing
2015-01-01
The principle of the X-ray detector which can simultaneously perform the measurement of the exposure rate and 2D (two-dimensional) distribution is described. A commercially available CMOS image sensor has been adopted as the key part to receive X-ray without any scintillators. The correlation between the pixel value (PV) and the absorbed exposure rate of X-ray is studied using the improved Elman neural network. Comparing the optimal adjustment process of the BP (Back Propagation) neural network and the improved Elman neural network, the neural network parameters are selected based on the fitting curve and the error curve. The experiments using the practical production data show that the proposed method achieves high accurate predictions to 10 −15 , which is consistent with the anticipated value. It is proven that it is possible to detect the exposure rate using the X-ray detector with the improved Elman algorithm for its advantages of fast converges and smooth error curve. - Highlights: • A method to measure the X-ray radiation with low cost and miniaturization. • A general CMOS image sensor is used to detect X-ray. • The system can measure exposure rate and 2D distribution simultaneously. • The Elman algorithm is adopted to improve the precision of the radiation detector
Directory of Open Access Journals (Sweden)
M. I. Fursanov
2014-01-01
Full Text Available This article reflects algorithmization of search methods of effective replacement of consumer transformers in distributed electrical networks. As any electrical equipment of power systems, power transformers have their own limited service duration, which is determined by natural processes of materials degradation and also by unexpected wear under different conditions of overload and overvoltage. According to the standards, adapted by in the Republic of Belarus, rated service life of power transformers is 25 years. But it can be situations that transformers should be better changed till this time – economically efficient. The possibility of such replacement is considered in order to increase efficiency of electrical network operation connected with its physical wear and aging.In this article the faults of early developed mathematical models of transformers replacement were discussed. Early such worked out transformers were not used. But in practice they can be replaced in one substation but they can be successfully used in other substations .Especially if there are limits of financial resources and the replacement needs more detail technical and economical basis.During the research the authors developed the efficient algorithm for determining of optimal location of transformers at substations of distributed electrical networks, based on search of the best solution from all sets of displacement in oriented graph. Suggested algorithm allows considerably reduce design time of optimal placement of transformers using a set of simplifications. The result of algorithm’s work is series displacement of transformers in networks, which allow obtain a great economic effect in comparison with replacement of single transformer.
International Nuclear Information System (INIS)
Viana, R.S.; Yoriyaz, H.; Santos, A.
2011-01-01
The Expectation-Maximization (E-M) algorithm is an iterative computational method for maximum likelihood (M-L) estimates, useful in a variety of incomplete-data problems. Due to its stochastic nature, one of the most relevant applications of E-M algorithm is the reconstruction of emission tomography images. In this paper, the statistical formulation of the E-M algorithm was applied to the in vivo spectrographic imaging of stable isotopes called Neutron Stimulated Emission Computed Tomography (NSECT). In the process of E-M algorithm iteration, the conditional probability distribution plays a very important role to achieve high quality image. This present work proposes an alternative methodology for the generation of the conditional probability distribution associated to the E-M reconstruction algorithm, using the Monte Carlo code MCNP5 and with the application of the reciprocity theorem. (author)
Energy Technology Data Exchange (ETDEWEB)
Viana, R.S.; Yoriyaz, H.; Santos, A., E-mail: rodrigossviana@gmail.com, E-mail: hyoriyaz@ipen.br, E-mail: asantos@ipen.br [Instituto de Pesquisas Energeticas e Nucleares (IPEN/CNEN-SP), Sao Paulo, SP (Brazil)
2011-07-01
The Expectation-Maximization (E-M) algorithm is an iterative computational method for maximum likelihood (M-L) estimates, useful in a variety of incomplete-data problems. Due to its stochastic nature, one of the most relevant applications of E-M algorithm is the reconstruction of emission tomography images. In this paper, the statistical formulation of the E-M algorithm was applied to the in vivo spectrographic imaging of stable isotopes called Neutron Stimulated Emission Computed Tomography (NSECT). In the process of E-M algorithm iteration, the conditional probability distribution plays a very important role to achieve high quality image. This present work proposes an alternative methodology for the generation of the conditional probability distribution associated to the E-M reconstruction algorithm, using the Monte Carlo code MCNP5 and with the application of the reciprocity theorem. (author)
Xu, Lei; Jeavons, Peter
2015-11-01
Leader election in anonymous rings and complete networks is a very practical problem in distributed computing. Previous algorithms for this problem are generally designed for a classical message passing model where complex messages are exchanged. However, the need to send and receive complex messages makes such algorithms less practical for some real applications. We present some simple synchronous algorithms for distributed leader election in anonymous rings and complete networks that are inspired by the development of the neural system of the fruit fly. Our leader election algorithms all assume that only one-bit messages are broadcast by nodes in the network and processors are only able to distinguish between silence and the arrival of one or more messages. These restrictions allow implementations to use a simpler message-passing architecture. Even with these harsh restrictions our algorithms are shown to achieve good time and message complexity both analytically and experimentally.
Directory of Open Access Journals (Sweden)
Rasoul Rajabpour
2017-01-01
Full Text Available Recent decades have witnessed growing applications of metaheuristic techniques as efficient tools for solving complex engineering problems. One such method is the JPSO algorithm. In this study, innovative modifications were made in the nature of the jump algorithm JPSO to make it capable of coping with graph-based solutions, which led to the development of a new algorithm called ‘G-JPSO’. The new algorithm was then used to solve the Fletcher-Powell optimal control problem and its application to optimal control of pumps in water distribution networks was evaluated. Optimal control of pumps consists in an optimum operation timetable (on and off for each of the pumps at the desired time interval. Maximum number of on and off positions for each pump was introduced into the objective function as a constraint such that not only would power consumption at each node be reduced but such problem requirements as the minimum pressure required at each node and minimum/maximum storage tank heights would be met. To determine the optimal operation of pumps, a model-based optimization-simulation algorithm was developed based on G-JPSO and JPSO algorithms. The model proposed by van Zyl was used to determine the optimal operation of the distribution network. Finally, the results obtained from the proposed algorithm were compared with those obtained from ant colony, genetic, and JPSO algorithms to show the robustness of the proposed algorithm in finding near-optimum solutions at reasonable computation costs.
Siragusa, Enrico; Haiminen, Niina; Utro, Filippo; Parida, Laxmi
2017-10-09
Computer simulations can be used to study population genetic methods, models and parameters, as well as to predict potential outcomes. For example, in plant populations, predicting the outcome of breeding operations can be studied using simulations. In-silico construction of populations with pre-specified characteristics is an important task in breeding optimization and other population genetic studies. We present two linear time Simulation using Best-fit Algorithms (SimBA) for two classes of problems where each co-fits two distributions: SimBA-LD fits linkage disequilibrium and minimum allele frequency distributions, while SimBA-hap fits founder-haplotype and polyploid allele dosage distributions. An incremental gap-filling version of previously introduced SimBA-LD is here demonstrated to accurately fit the target distributions, allowing efficient large scale simulations. SimBA-hap accuracy and efficiency is demonstrated by simulating tetraploid populations with varying numbers of founder haplotypes, we evaluate both a linear time greedy algoritm and an optimal solution based on mixed-integer programming. SimBA is available on http://researcher.watson.ibm.com/project/5669.
A Globally Convergent Parallel SSLE Algorithm for Inequality Constrained Optimization
Directory of Open Access Journals (Sweden)
Zhijun Luo
2014-01-01
Full Text Available A new parallel variable distribution algorithm based on interior point SSLE algorithm is proposed for solving inequality constrained optimization problems under the condition that the constraints are block-separable by the technology of sequential system of linear equation. Each iteration of this algorithm only needs to solve three systems of linear equations with the same coefficient matrix to obtain the descent direction. Furthermore, under certain conditions, the global convergence is achieved.
DEFF Research Database (Denmark)
Cao, Bin; Zhao, Jianwei; Yang, Po
2018-01-01
-objective evolutionary algorithms the Cooperative Coevolutionary Generalized Differential Evolution 3, the Cooperative Multi-objective Differential Evolution and the Nondominated Sorting Genetic Algorithm III, the proposed algorithm addresses the deployment optimization problem efficiently and effectively.......Using immune algorithms is generally a time-intensive process especially for problems with a large number of variables. In this paper, we propose a distributed parallel cooperative coevolutionary multi-objective large-scale immune algorithm that is implemented using the message passing interface...... (MPI). The proposed algorithm is composed of three layers: objective, group and individual layers. First, for each objective in the multi-objective problem to be addressed, a subpopulation is used for optimization, and an archive population is used to optimize all the objectives. Second, the large...
Directory of Open Access Journals (Sweden)
Li Ran
2017-01-01
Full Text Available Optimal allocation of generalized power sources in distribution network is researched. A simple index of voltage stability is put forward. Considering the investment and operation benefit, the stability of voltage and the pollution emissions of generalized power sources in distribution network, a multi-objective optimization planning model is established. A multi-objective particle swarm optimization algorithm is proposed to solve the optimal model. In order to improve the global search ability, the strategies of fast non-dominated sorting, elitism and crowding distance are adopted in this algorithm. Finally, tested the model and algorithm by IEEE-33 node system to find the best configuration of GP, the computed result shows that with the generalized power reasonable access to the active distribution network, the investment benefit and the voltage stability of the system is improved, and the proposed algorithm has better global search capability.
International Nuclear Information System (INIS)
He, Zhenzong; Qi, Hong; Yao, Yuchen; Ruan, Liming
2014-01-01
The Ant Colony Optimization algorithm based on the probability density function (PDF-ACO) is applied to estimate the bimodal aerosol particle size distribution (PSD). The direct problem is solved by the modified Anomalous Diffraction Approximation (ADA, as an approximation for optically large and soft spheres, i.e., χ⪢1 and |m−1|⪡1) and the Beer–Lambert law. First, a popular bimodal aerosol PSD and three other bimodal PSDs are retrieved in the dependent model by the multi-wavelength extinction technique. All the results reveal that the PDF-ACO algorithm can be used as an effective technique to investigate the bimodal PSD. Then, the Johnson's S B (J-S B ) function and the modified beta (M-β) function are employed as the general distribution function to retrieve the bimodal PSDs under the independent model. Finally, the J-S B and M-β functions are applied to recover actual measurement aerosol PSDs over Beijing and Shanghai obtained from the aerosol robotic network (AERONET). The numerical simulation and experimental results demonstrate that these two general functions, especially the J-S B function, can be used as a versatile distribution function to retrieve the bimodal aerosol PSD when no priori information about the PSD is available. - Highlights: • Bimodal PSDs are retrieved by ACO based on probability density function accurately. • J-S B and M-β functions can be used as the versatile function to recover bimodal PSDs. • Bimodal aerosol PSDs can be estimated by J-S B function more reasonably
Application of the distributed genetic algorithm for loading pattern optimization problems
International Nuclear Information System (INIS)
Hashimoto, Hiroshi; Yamamoto, Akio
2000-01-01
The distributed genetic algorithm (DGA) is applied for loading pattern optimization problems of the pressurized water reactors (PWR). Due to stiff nature of the loading pattern optimizations (e.g. multi-modality and non-linearity), stochastic methods like the simulated annealing or the genetic algorithm (GA) are widely applied for these problems. A basic concept of DGA is based on that of GA. However, DGA equally distributes candidates of solutions (i.e. loading patterns) to several independent 'islands' and evolves them in each island. Migrations of some candidates are performed among islands with a certain period. Since candidates of solutions independently evolve in each island with accepting different genes of migrants from other islands, premature convergence in the traditional GA can be prevented. Because many candidate loading patterns should be evaluated in one generation of GA or DGA, the parallelization in these calculations works efficiently. Parallel efficiency was measured using our optimization code and good load balance was attained even in a heterogeneous cluster environment due to dynamic distribution of the calculation load. The optimization code is based on the client/server architecture with the TCP/IP native socket and a client (optimization module) and calculation server modules communicate the objects of loading patterns each other. Throughout the sensitivity study on optimization parameters of DGA, a suitable set of the parameters for a test problem was identified. Finally, optimization capability of DGA and the traditional GA was compared in the test problem and DGA provided better optimization results than the traditional GA. (author)
Directory of Open Access Journals (Sweden)
Seyed Mehdi Miri
2014-07-01
Full Text Available The accidental or intentional entry of contaminants or self-deterioration of the water quality within the network itself can severely harm public health. Efficient water quality monitoring is one of the most important tools to guarantee a reliable potable water supply to consumers of drinking water distribution systems. Considering the high purchase, installation and maintenance cost of sensors in water distribution networks deploying two independent sensor networks within one distribution system is not only bounded by physical constraints but also is not a cost-effective approach. Therefore, need for combining different objectives and designing sensor network to simultaneity satisfying these objectives is felt. Sensors should comply with dual use benefits. Sensor locations and types should be integrated not only for achieving water security goals but also for accomplishing other water utility objectives, such as satisfying regulatory monitoring requirements or collecting information to solve water quality problems. In this study, a dual use vision for the sensor layout problem in the municipal water networks, is formulated and solved with the ant colony algorithm.
Wang, Li; Li, Feng; Xing, Jian
2017-10-01
In this paper, a hybrid artificial bee colony (ABC) algorithm and pattern search (PS) method is proposed and applied for recovery of particle size distribution (PSD) from spectral extinction data. To be more useful and practical, size distribution function is modelled as the general Johnson's ? function that can overcome the difficulty of not knowing the exact type beforehand encountered in many real circumstances. The proposed hybrid algorithm is evaluated through simulated examples involving unimodal, bimodal and trimodal PSDs with different widths and mean particle diameters. For comparison, all examples are additionally validated by the single ABC algorithm. In addition, the performance of the proposed algorithm is further tested by actual extinction measurements with real standard polystyrene samples immersed in water. Simulation and experimental results illustrate that the hybrid algorithm can be used as an effective technique to retrieve the PSDs with high reliability and accuracy. Compared with the single ABC algorithm, our proposed algorithm can produce more accurate and robust inversion results while taking almost comparative CPU time over ABC algorithm alone. The superiority of ABC and PS hybridization strategy in terms of reaching a better balance of estimation accuracy and computation effort increases its potentials as an excellent inversion technique for reliable and efficient actual measurement of PSD.
International Nuclear Information System (INIS)
Liu, L; Yuan, F G
2008-01-01
Wireless structural health monitoring (SHM) systems have emerged as a promising technology for robust and cost-effective structural monitoring. However, the applications of wireless sensors on active diagnosis for structural health monitoring (SHM) have not been extensively investigated. Due to limited energy sources, battery-powered wireless sensors can only perform limited functions and are expected to operate at a low duty cycle. Conventional designs are not suitable for sensing high frequency signals, e.g. in the ultrasonic frequency range. More importantly, algorithms to detect structural damage with a vast amount of data usually require considerable processing and communication time and result in unaffordable power consumption for wireless sensors. In this study, an energy-efficient wireless sensor for supporting high frequency signals and a distributed damage localization algorithm for plate-like structures are proposed, discussed and validated to supplement recent advances made for active sensing-based SHM. First, the power consumption of a wireless sensor is discussed and identified. Then the design of a wireless sensor for active diagnosis using piezoelectric sensors is introduced. The newly developed wireless sensor utilizes an optimized combination of field programmable gate array (FPGA) and conventional microcontroller to address the tradeoff between power consumption and speed requirement. The proposed damage localization algorithm, based on an energy decay model, enables wireless sensors to be practically used in active diagnosis. The power consumption for data communication can be minimized while the power budget for data processing can still be affordable for a battery-powered wireless sensor. The Levenberg–Marquardt method is employed in a mains-powered sensor node or PC to locate damage. Experimental results and discussion on the improvement of power efficiency are given
Directory of Open Access Journals (Sweden)
Alagan Anpalagan
2013-04-01
Full Text Available Due to the rapid increase in the usage and demand of wireless sensor networks (WSN, the limited frequency spectrum available for WSN applications will be extremely crowded in the near future. More sensor devices also mean more recharging/replacement of batteries, which will cause significant impact on the global carbon footprint. In this paper, we propose a relay-assisted cognitive radio sensor network (CRSN that allocates communication resources in an environmentally friendly manner. We use shared band amplify and forward relaying for cooperative communication in the proposed CRSN. We present a multi-objective optimization architecture for resource allocation in a green cooperative cognitive radio sensor network (GC-CRSN. The proposed multi-objective framework jointly performs relay assignment and power allocation in GC-CRSN, while optimizing two conflicting objectives. The first objective is to maximize the total throughput, and the second objective is to minimize the total transmission power of CRSN. The proposed relay assignment and power allocation problem is a non-convex mixed-integer non-linear optimization problem (NC-MINLP, which is generally non-deterministic polynomial-time (NP-hard. We introduce a hybrid heuristic algorithm for this problem. The hybrid heuristic includes an estimation-of-distribution algorithm (EDA for performing power allocation and iterative greedy schemes for constraint satisfaction and relay assignment. We analyze the throughput and power consumption tradeoff in GC-CRSN. A detailed analysis of the performance of the proposed algorithm is presented with the simulation results.
Constrained treatment planning using sequential beam selection
International Nuclear Information System (INIS)
Woudstra, E.; Storchi, P.R.M.
2000-01-01
In this paper an algorithm is described for automated treatment plan generation. The algorithm aims at delivery of the prescribed dose to the target volume without violation of constraints for target, organs at risk and the surrounding normal tissue. Pre-calculated dose distributions for all candidate orientations are used as input. Treatment beams are selected in a sequential way. A score function designed for beam selection is used for the simultaneous selection of beam orientations and weights. In order to determine the optimum choice for the orientation and the corresponding weight of each new beam, the score function is first redefined to account for the dose distribution of the previously selected beams. Addition of more beams to the plan is stopped when the target dose is reached or when no additional dose can be delivered without violating a constraint. In the latter case the score function is modified by importance factor changes to enforce better sparing of the organ with the limiting constraint and the algorithm is run again. (author)
Double-layer evolutionary algorithm for distributed optimization of particle detection on the Grid
International Nuclear Information System (INIS)
Padée, Adam; Zaremba, Krzysztof; Kurek, Krzysztof
2013-01-01
Reconstruction of particle tracks from information collected by position-sensitive detectors is an important procedure in HEP experiments. It is usually controlled by a set of numerical parameters which have to be manually optimized. This paper proposes an automatic approach to this task by utilizing evolutionary algorithm (EA) operating on both real-valued and binary representations. Because of computational complexity of the task a special distributed architecture of the algorithm is proposed, designed to be run in grid environment. It is two-level hierarchical hybrid utilizing asynchronous master-slave EA on the level of clusters and island model EA on the level of the grid. The technical aspects of usage of production grid infrastructure are covered, including communication protocols on both levels. The paper deals also with the problem of heterogeneity of the resources, presenting efficiency tests on a benchmark function. These tests confirm that even relatively small islands (clusters) can be beneficial to the optimization process when connected to the larger ones. Finally a real-life usage example is presented, which is an optimization of track reconstruction in Large Angle Spectrometer of NA-58 COMPASS experiment held at CERN, using a sample of Monte Carlo simulated data. The overall reconstruction efficiency gain, achieved by the proposed method, is more than 4%, compared to the manually optimized parameters
An Efficient MapReduce-Based Parallel Clustering Algorithm for Distributed Traffic Subarea Division
Directory of Open Access Journals (Sweden)
Dawen Xia
2015-01-01
Full Text Available Traffic subarea division is vital for traffic system management and traffic network analysis in intelligent transportation systems (ITSs. Since existing methods may not be suitable for big traffic data processing, this paper presents a MapReduce-based Parallel Three-Phase K-Means (Par3PKM algorithm for solving traffic subarea division problem on a widely adopted Hadoop distributed computing platform. Specifically, we first modify the distance metric and initialization strategy of K-Means and then employ a MapReduce paradigm to redesign the optimized K-Means algorithm for parallel clustering of large-scale taxi trajectories. Moreover, we propose a boundary identifying method to connect the borders of clustering results for each cluster. Finally, we divide traffic subarea of Beijing based on real-world trajectory data sets generated by 12,000 taxis in a period of one month using the proposed approach. Experimental evaluation results indicate that when compared with K-Means, Par2PK-Means, and ParCLARA, Par3PKM achieves higher efficiency, more accuracy, and better scalability and can effectively divide traffic subarea with big taxi trajectory data.
Directory of Open Access Journals (Sweden)
Pedro L. Iglesias
2007-01-01
Full Text Available The Genetic Algorithms (GAs are a technique of optimization used for water distribution networks design. This work has been made with a modified pseudo genetic algorithm (PGA, whose main variation with a classical GA is a change in the codification of the chromosomes, which is made of numerical form instead of the binary codification. This variation entails a series of special characteristics in the codification and in the definition of the operations of mutation and crossover. Initially, the work displays the results of the PGA on a water network studied in the literature. The results show the kindness of the method. Also is made a statistical analysis of the obtained solutions. This analysis allows verifying the values of mutation and crossing probability more suitable for the proposed method. Finally, in the study of the analyzed water supply networks the concept of reliability in introduced. This concept is essential to understand the validity of the obtained results. The second part, starting with values optimized for the probability of crossing and mutation, the influence of the population size is analyzed in the final solutions on the network of Hanoi, widely studied in the bibliography. The aim is to find the most suitable configuration of the problem, so that good solutions are obtained in the less time.
Directory of Open Access Journals (Sweden)
Pedro L. Iglesias
2007-12-01
Full Text Available The Genetic Algorithms (GAs are a technique of optimization used for water distribution networks design. This work has been made with a modified pseudo genetic algorithm (PGA, whose main variation with a classical GA is a change in the codification of the chromosomes, which is made of numerical form instead of the binary codification. This variation entails a series of special characteristics in the codification and in the definition of the operations of mutation and crossover. Initially, the work displays the results of the PGA on a water network studied in the literature. The results show the kindness of the method. Also is made a statistical analysis of the obtained solutions. This analysis allows verifying the values of mutation and crossing probability more suitable for the proposed method. Finally, in the study of the analyzed water supply networks the concept of reliability in introduced. This concept is essential to understand the validity of the obtained results. The second part, starting with values optimized for the probability of crossing and mutation, the influence of the population size is analyzed in the final solutions on the network of Hanoi, widely studied in the bibliography. The aim is to find the most suitable configuration of the problem, so that good solutions are obtained in the less time.
International Nuclear Information System (INIS)
Martin-del-Campo, Cecilia; Francois, Juan Luis; Carmona, Roberto; Oropeza, Ivonne P.
2007-01-01
An optimization methodology based on the Genetic Algorithms (GA) method was developed for the design of radial enrichment and gadolinia distributions for boiling water reactor (BWR) fuel lattices. The optimization algorithm was linked to the HELIOS code to evaluate the neutronic parameters included in the objective function. The goal is to search for a fuel lattice with the lowest average enrichment, which satisfy a reactivity target, a local power peaking factor (PPF), lower than a limit value, and an average gadolinia concentration target. The methodology was applied to the design of a 10 x 10 fuel lattice, which can be used in fuel assemblies currently used in the two BWRs operating at Mexico. The optimization process showed an excellent performance because it found forty lattice designs in which the worst one has a better neutronic performance than the reference lattice design. The main contribution of this study is the development of an efficient procedure for BWR fuel lattice design, using GA with an objective function (OF) which saves computing time because it does not require lattice burnup calculations
Distributed control software of high-performance control-loop algorithm
Blanc, D
1999-01-01
The majority of industrial cooling and ventilation plants require the control of complex processes. All these processes are highly important for the operation of the machines. The stability and reliability of these processes are leading factors identifying the quality of the service provided. The control system architecture and software structure, as well, are required to have high dynamical performance and robust behaviour. The intelligent systems based on PID or RST controllers are used for their high level of stability and accuracy. The design and tuning of these complex controllers require the dynamic model of the plant to be known (generally obtained by identification) and the desired performance of the various control loops to be specified for achieving good performances. The concept of having a distributed control algorithm software provides full automation facilities with well-adapted functionality and good performances, giving methodology, means and tools to master the dynamic process optimization an...
Directory of Open Access Journals (Sweden)
Xite Wang
2017-01-01
Full Text Available Outlier detection is an important data mining task, whose target is to find the abnormal or atypical objects from a given dataset. The techniques for detecting outliers have a lot of applications, such as credit card fraud detection and environment monitoring. Our previous work proposed the Cluster-Based (CB outlier and gave a centralized method using unsupervised extreme learning machines to compute CB outliers. In this paper, we propose a new distributed algorithm for the CB outlier detection (DACB. On the master node, we collect a small number of points from the slave nodes to obtain a threshold. On each slave node, we design a new filtering method that can use the threshold to efficiently speed up the computation. Furthermore, we also propose a ranking method to optimize the order of cluster scanning. At last, the effectiveness and efficiency of the proposed approaches are verified through a plenty of simulation experiments.
Fuzzy pool balance: An algorithm to achieve a two dimensional balance in distribute storage systems
International Nuclear Information System (INIS)
Wu, Wenjing; Chen, Gang
2014-01-01
The limitation of scheduling modules and the gradual addition of disk pools in distributed storage systems often result in imbalances among their disk pools in terms of both disk usage and file count. This can cause various problems to the storage system such as single point of failure, low system throughput and imbalanced resource utilization and system loads. An algorithm named Fuzzy Pool Balance (FPB) is proposed here to solve this problem. The input of FPB is the current file distribution among disk pools and the output is a file migration plan indicating what files are to be migrated to which pools. FPB uses an array to classify the files by their sizes. The file classification array is dynamically calculated with a defined threshold named T max that defines the allowed pool disk usage deviations. File classification is the basis of file migration. FPB also defines the Immigration Pool (IP) and Emigration Pool (EP) according to the pool disk usage and File Quantity Ratio (FQR) that indicates the percentage of each category of files in each disk pool, so files with higher FQR in an EP will be migrated to IP(s) with a lower FQR of this file category. To verify this algorithm, we implemented FPB on an ATLAS Tier2 dCache production system. The results show that FPB can achieve a very good balance in both free space and file counts, and adjusting the threshold value T max and the correction factor to the average FQR can achieve a tradeoff between free space and file count.
International Nuclear Information System (INIS)
Perles, L.A.; Chinellato, C.D.; Rocha, J.R.O.
2001-01-01
In this paper has been presented a modification of a algorithm for three-dimensional (3D) radiation dose distribution in heterogeneous media by convolutions. This modification has maintained good accordance between calculated and simulated data in EGS4 code. The results of algorithm have been compared with commercial program PLATO, where have been noticed inconsistency for equivalent density regions in a muscle-lung-muscle interface system
International Nuclear Information System (INIS)
Koch, Nicholas C; Newhauser, Wayne D
2010-01-01
Proton beam radiotherapy is an effective and non-invasive treatment for uveal melanoma. Recent research efforts have focused on improving the dosimetric accuracy of treatment planning and overcoming the present limitation of relative analytical dose calculations. Monte Carlo algorithms have been shown to accurately predict dose per monitor unit (D/MU) values, but this has yet to be shown for analytical algorithms dedicated to ocular proton therapy, which are typically less computationally expensive than Monte Carlo algorithms. The objective of this study was to determine if an analytical method could predict absolute dose distributions and D/MU values for a variety of treatment fields like those used in ocular proton therapy. To accomplish this objective, we used a previously validated Monte Carlo model of an ocular nozzle to develop an analytical algorithm to predict three-dimensional distributions of D/MU values from pristine Bragg peaks and therapeutically useful spread-out Bragg peaks (SOBPs). Results demonstrated generally good agreement between the analytical and Monte Carlo absolute dose calculations. While agreement in the proximal region decreased for beams with less penetrating Bragg peaks compared with the open-beam condition, the difference was shown to be largely attributable to edge-scattered protons. A method for including this effect in any future analytical algorithm was proposed. Comparisons of D/MU values showed typical agreement to within 0.5%. We conclude that analytical algorithms can be employed to accurately predict absolute proton dose distributions delivered by an ocular nozzle.
Communication complexity of distributed computing and a parallel algorithm for polynomial roots
International Nuclear Information System (INIS)
Tiwari, P.
1986-01-01
The first part of this thesis begins with a discussion of the minimum communication requirements in some distributed networks. The main result is a general technique for determining lower bounds on the communication complexity of problems on various distributed computer networks. This general technique is derived by simulating the general network by a linear array and then using a lower bound on the communication complexity of the problem on the linear array. Applications of this technique yield nontrivial optimal or near-optimal lower bounds on the communication complexity of distinctness, ranking, uniqueness, merging, and triangle detection on a ring, a mesh, and a complete binary tree of processors. A technique similar to the one used in proving the above results, yields interesting graph theoretic results concerning decomposition of a graph into complete bipartite subgraphs. The second part of the this is devoted to the design of a fast parallel algorithm for determining all roots of a polynomial. Given a polynomial rho(z) of degree n with m bit integer coefficients and an integer μ, the author considers the problem of determining all its roots with error less than 2/sup -μ/. It is shown that this problem is in the class NC if rho(z) has all real roots
Genetic algorithms for the optimization of pipeline systems for liquid distribution (2)
International Nuclear Information System (INIS)
Narvaez, Paulo Cesar; Galeano, Haiver
2004-01-01
This is the second of two articles presenting a Genetic Algorithm (GA) to obtain an optimal design, from an economical and operational point of view, of a pipeline system for the distribution of liquids, based on criteria such as complying with the laws of preservation of mass and energy, volume of flow requirements in the points of consumption where pressure is known, restriction in pressure value in those points of the system where it is unknown as well as in the velocity which must be under the erosion limit. In this article the traditional techniques for designing a GA in this type of problems are combined with some ideas that have not been applied to this field previously. The proposed GA allows for the sizing of liquid distribution systems that include pipelines, nodes for consumption and provision, tanks, pumping equipment, nozzles, control valves and accessories. The first article of this series (Galeano, 2003), presents the different formulations found in literature for the design of networks through optimization techniques and formulates mathematically, the optimization problem. In this article, the characteristics of the GA are specified and it is applied to solve the Alperovits and Shamir (1977) network and for a fireproof network, which allowed testing some of the characteristics of the model that are not found in the literature, such as the possibility of including pumping equipment, aspersion nozzles and accessories. In addition, the contribution of the components and sensitivity are analyzed in order to investigate some characteristics and parameters of the implemented GA
Caballero-Águila, R.; Hermoso-Carazo, A.; Linares-Pérez, J.
2017-07-01
This paper studies the distributed fusion estimation problem from multisensor measured outputs perturbed by correlated noises and uncertainties modelled by random parameter matrices. Each sensor transmits its outputs to a local processor over a packet-erasure channel and, consequently, random losses may occur during transmission. Different white sequences of Bernoulli variables are introduced to model the transmission losses. For the estimation, each lost output is replaced by its estimator based on the information received previously, and only the covariances of the processes involved are used, without requiring the signal evolution model. First, a recursive algorithm for the local least-squares filters is derived by using an innovation approach. Then, the cross-correlation matrices between any two local filters is obtained. Finally, the distributed fusion filter weighted by matrices is obtained from the local filters by applying the least-squares criterion. The performance of the estimators and the influence of both sensor uncertainties and transmission losses on the estimation accuracy are analysed in a numerical example.
Energy Technology Data Exchange (ETDEWEB)
Jenq, B.C.
1986-01-01
The performance evaluation of integrated concurrency-control and recovery mechanisms for distributed data base systems is studied using a distributed testbed system. In addition, a queueing network model was developed to analyze the two phase locking scheme in the distributed testbed system. The combination of testbed measurement and analytical modeling provides an effective tool for understanding the performance of integrated concurrency control and recovery algorithms in distributed database systems. The design and implementation of the distributed testbed system, CARAT, are presented. The concurrency control and recovery algorithms implemented in CARAT include: a two phase locking scheme with distributed deadlock detection, a distributed version of optimistic approach, before-image and after-image journaling mechanisms for transaction recovery, and a two-phase commit protocol. Many performance measurements were conducted using a variety of workloads. A queueing network model is developed to analyze the performance of the CARAT system using the two-phase locking scheme with before-image journaling. The combination of testbed measurements and analytical modeling provides significant improvements in understanding the performance impacts of the concurrency control and recovery algorithms in distributed database systems.
Directory of Open Access Journals (Sweden)
R. Rajaram
2015-11-01
Full Text Available Network reconfiguration which is constrained non linear optimization problem has been solved for loss minimization, load balancing, etc. for last two decades using various heuristic search evolutionary algorithms like binary particle swarm optimization, neuro-fuzzy techniques, etc. The contribution of this paper lies in considering distributed generation which are smaller power sources like solar photovoltaic cells or wind turbines connected in the customer roof top. This new connection in the radial network has made unidirectional current flow to become bidirectional there by increasing the efficiency but sometimes reducing stability of the system. Modified plant growth simulation algorithm has been applied here successfully to minimize real power loss because it does not require barrier factors or cross over rates because the objectives and constraints are dealt separately. The main advantage of this algorithm is continuous guiding search along with changing objective function because power from distributed generation is continuously varying so this can be applied for real time applications with required modifications. This algorithm here is tested for a standard 33 bus radial distribution system for loss minimization and test results here shows that this algorithm is efficient and suitable for real time applications.
International Nuclear Information System (INIS)
Miyatake, Aya; Nishio, Teiji
2013-01-01
Purpose: Recently, much research on imaging the clinical proton-irradiated volume using positron emitter nuclei based on target nuclear fragment reaction has been carried out. The purpose of this study is to develop an activity pencil beam (APB) algorithm for a simulation system for proton-activated positron-emitting imaging in clinical proton therapy using spread-out Bragg peak (SOBP) beams.Methods: The target nuclei of activity distribution calculations are 12 C nuclei, 16 O nuclei, and 40 Ca nuclei, which are the main elements in a human body. Depth activity distributions with SOBP beam irradiations were obtained from the material information of ridge filter (RF) and depth activity distributions of compounds of the three target nuclei measured by BOLPs-RGp (beam ON-LINE PET system mounted on a rotating gantry port) with mono-energetic Bragg peak (MONO) beam irradiations. The calculated data of depth activity distributions with SOBP beam irradiations were sorted in terms of kind of nucleus, energy of proton beam, SOBP width, and thickness of fine degrader (FD), which were verified. The calculated depth activity distributions with SOBP beam irradiations were compared with the measured ones. APB kernels were made from the calculated depth activity distributions with SOBP beam irradiations to construct a simulation system using the APB algorithm for SOBP beams.Results: The depth activity distributions were prepared using the material information of RF and the measured depth activity distributions with MONO beam irradiations for clinical therapy using SOBP beams. With the SOBP width widening, the distal fall-offs of depth activity distributions and the difference from the depth dose distributions were large. The shapes of the calculated depth activity distributions nearly agreed with those of the measured ones upon comparison between the two. The APB kernels of SOBP beams were prepared by making use of the data on depth activity distributions with SOBP beam
An Efficient System Based On Closed Sequential Patterns for Web Recommendations
Utpala Niranjan; R.B.V. Subramanyam; V-Khana
2010-01-01
Sequential pattern mining, since its introduction has received considerable attention among the researchers with broad applications. The sequential pattern algorithms generally face problems when mining long sequential patterns or while using very low support threshold. One possible solution of such problems is by mining the closed sequential patterns, which is a condensed representation of sequential patterns. Recently, several researchers have utilized the sequential pattern discovery for d...
Wong, Ling Ai; Shareef, Hussain; Mohamed, Azah; Ibrahim, Ahmad Asrul
2014-01-01
This paper presents the application of enhanced opposition-based firefly algorithm in obtaining the optimal battery energy storage systems (BESS) sizing in photovoltaic generation integrated radial distribution network in order to mitigate the voltage rise problem. Initially, the performance of the original firefly algorithm is enhanced by utilizing the opposition-based learning and introducing inertia weight. After evaluating the performance of the enhanced opposition-based firefly algorithm (EOFA) with fifteen benchmark functions, it is then adopted to determine the optimal size for BESS. Two optimization processes are conducted where the first optimization aims to obtain the optimal battery output power on hourly basis and the second optimization aims to obtain the optimal BESS capacity by considering the state of charge constraint of BESS. The effectiveness of the proposed method is validated by applying the algorithm to the 69-bus distribution system and by comparing the performance of EOFA with conventional firefly algorithm and gravitational search algorithm. Results show that EOFA has the best performance comparatively in terms of mitigating the voltage rise problem.
Directory of Open Access Journals (Sweden)
Ling Ai Wong
2014-01-01
Full Text Available This paper presents the application of enhanced opposition-based firefly algorithm in obtaining the optimal battery energy storage systems (BESS sizing in photovoltaic generation integrated radial distribution network in order to mitigate the voltage rise problem. Initially, the performance of the original firefly algorithm is enhanced by utilizing the opposition-based learning and introducing inertia weight. After evaluating the performance of the enhanced opposition-based firefly algorithm (EOFA with fifteen benchmark functions, it is then adopted to determine the optimal size for BESS. Two optimization processes are conducted where the first optimization aims to obtain the optimal battery output power on hourly basis and the second optimization aims to obtain the optimal BESS capacity by considering the state of charge constraint of BESS. The effectiveness of the proposed method is validated by applying the algorithm to the 69-bus distribution system and by comparing the performance of EOFA with conventional firefly algorithm and gravitational search algorithm. Results show that EOFA has the best performance comparatively in terms of mitigating the voltage rise problem.
Pronk, Sander; Pouya, Iman; Lundborg, Magnus; Rotskoff, Grant; Wesén, Björn; Kasson, Peter M; Lindahl, Erik
2015-06-09
Computational chemistry and other simulation fields are critically dependent on computing resources, but few problems scale efficiently to the hundreds of thousands of processors available in current supercomputers-particularly for molecular dynamics. This has turned into a bottleneck as new hardware generations primarily provide more processing units rather than making individual units much faster, which simulation applications are addressing by increasingly focusing on sampling with algorithms such as free-energy perturbation, Markov state modeling, metadynamics, or milestoning. All these rely on combining results from multiple simulations into a single observation. They are potentially powerful approaches that aim to predict experimental observables directly, but this comes at the expense of added complexity in selecting sampling strategies and keeping track of dozens to thousands of simulations and their dependencies. Here, we describe how the distributed execution framework Copernicus allows the expression of such algorithms in generic workflows: dataflow programs. Because dataflow algorithms explicitly state dependencies of each constituent part, algorithms only need to be described on conceptual level, after which the execution is maximally parallel. The fully automated execution facilitates the optimization of these algorithms with adaptive sampling, where undersampled regions are automatically detected and targeted without user intervention. We show how several such algorithms can be formulated for computational chemistry problems, and how they are executed efficiently with many loosely coupled simulations using either distributed or parallel resources with Copernicus.
Directory of Open Access Journals (Sweden)
Martin Straka
2015-03-01
Full Text Available Reducing costs forces of companies to look for reserves also in field of management, support and implementation of material flow. This is connected with the optimization of costs for product distribution, which forms a significant part of the total cost of the company. In practice, it happens that making plans for distribution of materials within a distribution space is left solely to the implementers of distribution and it drivers of vehicles. As a result is uneconomic material distribution and unnecessary cost increases to his distribution. The objective of the paper is to propose a methodology for optimization of transport planning from DC Prešov to the individual Tesco units within the region that comes under this distribution centre. The methodology is based on the Clarke and Wright saving algorithm.
The HTM Spatial Pooler—A Neocortical Algorithm for Online Sparse Distributed Coding
Directory of Open Access Journals (Sweden)
Yuwei Cui
2017-11-01
Full Text Available Hierarchical temporal memory (HTM provides a theoretical framework that models several key computational principles of the neocortex. In this paper, we analyze an important component of HTM, the HTM spatial pooler (SP. The SP models how neurons learn feedforward connections and form efficient representations of the input. It converts arbitrary binary input patterns into sparse distributed representations (SDRs using a combination of competitive Hebbian learning rules and homeostatic excitability control. We describe a number of key properties of the SP, including fast adaptation to changing input statistics, improved noise robustness through learning, efficient use of cells, and robustness to cell death. In order to quantify these properties we develop a set of metrics that can be directly computed from the SP outputs. We show how the properties are met using these metrics and targeted artificial simulations. We then demonstrate the value of the SP in a complete end-to-end real-world HTM system. We discuss the relationship with neuroscience and previous studies of sparse coding. The HTM spatial pooler represents a neurally inspired algorithm for learning sparse representations from noisy data streams in an online fashion.
The HTM Spatial Pooler-A Neocortical Algorithm for Online Sparse Distributed Coding.
Cui, Yuwei; Ahmad, Subutai; Hawkins, Jeff
2017-01-01
Hierarchical temporal memory (HTM) provides a theoretical framework that models several key computational principles of the neocortex. In this paper, we analyze an important component of HTM, the HTM spatial pooler (SP). The SP models how neurons learn feedforward connections and form efficient representations of the input. It converts arbitrary binary input patterns into sparse distributed representations (SDRs) using a combination of competitive Hebbian learning rules and homeostatic excitability control. We describe a number of key properties of the SP, including fast adaptation to changing input statistics, improved noise robustness through learning, efficient use of cells, and robustness to cell death. In order to quantify these properties we develop a set of metrics that can be directly computed from the SP outputs. We show how the properties are met using these metrics and targeted artificial simulations. We then demonstrate the value of the SP in a complete end-to-end real-world HTM system. We discuss the relationship with neuroscience and previous studies of sparse coding. The HTM spatial pooler represents a neurally inspired algorithm for learning sparse representations from noisy data streams in an online fashion.
Directory of Open Access Journals (Sweden)
Fernando Cervantes-Sanchez
2016-01-01
Full Text Available This paper presents a novel method for improving the training step of the single-scale Gabor filters by using the Boltzmann univariate marginal distribution algorithm (BUMDA in X-ray angiograms. Since the single-scale Gabor filters (SSG are governed by three parameters, the optimal selection of the SSG parameters is highly desirable in order to maximize the detection performance of coronary arteries while reducing the computational time. To obtain the best set of parameters for the SSG, the area (Az under the receiver operating characteristic curve is used as fitness function. Moreover, to classify vessel and nonvessel pixels from the Gabor filter response, the interclass variance thresholding method has been adopted. The experimental results using the proposed method obtained the highest detection rate with Az=0.9502 over a training set of 40 images and Az=0.9583 with a test set of 40 images. In addition, the experimental results of vessel segmentation provided an accuracy of 0.944 with the test set of angiograms.
Directory of Open Access Journals (Sweden)
Kourosh Behzadian
2008-03-01
Full Text Available In this paper, a novel multiobjective optimization model is presented for selecting optimal locations in the water distribution network (WDN with the aim of installing pressure loggers. The pressure data collected at optimal locations will be used later on in the calibration of the proposed WDN model. Objective functions consist of maximization of calibrated model prediction accuracy and minimization of the total cost for sampling design. In order to decrease the model run time, an optimization model has been developed using multiobjective genetic algorithm and adaptive neural network (MOGA-ANN. Neural networks (NNs are initially trained after a number of initial GA generations and periodically retrained and updated after generation of a specified number of full model-analyzed solutions. Trained NNs are replaced with the fitness evaluation of some chromosomes within the GA progress. Using cache prevents objective function evaluation of repetitive chromosomes within GA. Optimal solutions are obtained through pareto-optimal front with respect to the two objective functions. Results show that jointing NNs in MOGA for approximating portions of chromosomes’ fitness in each generation leads to considerable savings in model run time and can be promising for reducing run-time in optimization models with significant computational effort.
Optimal sensor placement for leak location in water distribution networks using genetic algorithms.
Casillas, Myrna V; Puig, Vicenç; Garza-Castañón, Luis E; Rosich, Albert
2013-11-04
This paper proposes a new sensor placement approach for leak location in water distribution networks (WDNs). The sensor placement problem is formulated as an integer optimization problem. The optimization criterion consists in minimizing the number of non-isolable leaks according to the isolability criteria introduced. Because of the large size and non-linear integer nature of the resulting optimization problem, genetic algorithms (GAs) are used as the solution approach. The obtained results are compared with a semi-exhaustive search method with higher computational effort, proving that GA allows one to find near-optimal solutions with less computational load. Moreover, three ways of increasing the robustness of the GA-based sensor placement method have been proposed using a time horizon analysis, a distance-based scoring and considering different leaks sizes. A great advantage of the proposed methodology is that it does not depend on the isolation method chosen by the user, as long as it is based on leak sensitivity analysis. Experiments in two networks allow us to evaluate the performance of the proposed approach.
Optimal Sensor Placement for Leak Location in Water Distribution Networks Using Genetic Algorithms
Directory of Open Access Journals (Sweden)
Luis E. Garza-Castañón
2013-11-01
Full Text Available This paper proposes a new sensor placement approach for leak location in water distribution networks (WDNs. The sensor placement problem is formulated as an integer optimization problem. The optimization criterion consists in minimizing the number of non-isolable leaks according to the isolability criteria introduced. Because of the large size and non-linear integer nature of the resulting optimization problem, genetic algorithms (GAs are used as the solution approach. The obtained results are compared with a semi-exhaustive search method with higher computational effort, proving that GA allows one to find near-optimal solutions with less computational load. Moreover, three ways of increasing the robustness of the GA-based sensor placement method have been proposed using a time horizon analysis, a distance-based scoring and considering different leaks sizes. A great advantage of the proposed methodology is that it does not depend on the isolation method chosen by the user, as long as it is based on leak sensitivity analysis. Experiments in two networks allow us to evaluate the performance of the proposed approach.
Optimal Sensor Placement for Leak Location in Water Distribution Networks Using Genetic Algorithms
Casillas, Myrna V.; Puig, Vicenç; Garza-Castañón, Luis E.; Rosich, Albert
2013-01-01
This paper proposes a new sensor placement approach for leak location in water distribution networks (WDNs). The sensor placement problem is formulated as an integer optimization problem. The optimization criterion consists in minimizing the number of non-isolable leaks according to the isolability criteria introduced. Because of the large size and non-linear integer nature of the resulting optimization problem, genetic algorithms (GAs) are used as the solution approach. The obtained results are compared with a semi-exhaustive search method with higher computational effort, proving that GA allows one to find near-optimal solutions with less computational load. Moreover, three ways of increasing the robustness of the GA-based sensor placement method have been proposed using a time horizon analysis, a distance-based scoring and considering different leaks sizes. A great advantage of the proposed methodology is that it does not depend on the isolation method chosen by the user, as long as it is based on leak sensitivity analysis. Experiments in two networks allow us to evaluate the performance of the proposed approach. PMID:24193099
Directory of Open Access Journals (Sweden)
Junguo Hu
Full Text Available Soil respiration inherently shows strong spatial variability. It is difficult to obtain an accurate characterization of soil respiration with an insufficient number of monitoring points. However, it is expensive and cumbersome to deploy many sensors. To solve this problem, we proposed employing the Bayesian Maximum Entropy (BME algorithm, using soil temperature as auxiliary information, to study the spatial distribution of soil respiration. The BME algorithm used the soft data (auxiliary information effectively to improve the estimation accuracy of the spatiotemporal distribution of soil respiration. Based on the functional relationship between soil temperature and soil respiration, the BME algorithm satisfactorily integrated soil temperature data into said spatial distribution. As a means of comparison, we also applied the Ordinary Kriging (OK and Co-Kriging (Co-OK methods. The results indicated that the root mean squared errors (RMSEs and absolute values of bias for both Day 1 and Day 2 were the lowest for the BME method, thus demonstrating its higher estimation accuracy. Further, we compared the performance of the BME algorithm coupled with auxiliary information, namely soil temperature data, and the OK method without auxiliary information in the same study area for 9, 21, and 37 sampled points. The results showed that the RMSEs for the BME algorithm (0.972 and 1.193 were less than those for the OK method (1.146 and 1.539 when the number of sampled points was 9 and 37, respectively. This indicates that the former method using auxiliary information could reduce the required number of sampling points for studying spatial distribution of soil respiration. Thus, the BME algorithm, coupled with soil temperature data, can not only improve the accuracy of soil respiration spatial interpolation but can also reduce the number of sampling points.
Hu, Junguo; Zhou, Jian; Zhou, Guomo; Luo, Yiqi; Xu, Xiaojun; Li, Pingheng; Liang, Junyi
2016-01-01
Soil respiration inherently shows strong spatial variability. It is difficult to obtain an accurate characterization of soil respiration with an insufficient number of monitoring points. However, it is expensive and cumbersome to deploy many sensors. To solve this problem, we proposed employing the Bayesian Maximum Entropy (BME) algorithm, using soil temperature as auxiliary information, to study the spatial distribution of soil respiration. The BME algorithm used the soft data (auxiliary information) effectively to improve the estimation accuracy of the spatiotemporal distribution of soil respiration. Based on the functional relationship between soil temperature and soil respiration, the BME algorithm satisfactorily integrated soil temperature data into said spatial distribution. As a means of comparison, we also applied the Ordinary Kriging (OK) and Co-Kriging (Co-OK) methods. The results indicated that the root mean squared errors (RMSEs) and absolute values of bias for both Day 1 and Day 2 were the lowest for the BME method, thus demonstrating its higher estimation accuracy. Further, we compared the performance of the BME algorithm coupled with auxiliary information, namely soil temperature data, and the OK method without auxiliary information in the same study area for 9, 21, and 37 sampled points. The results showed that the RMSEs for the BME algorithm (0.972 and 1.193) were less than those for the OK method (1.146 and 1.539) when the number of sampled points was 9 and 37, respectively. This indicates that the former method using auxiliary information could reduce the required number of sampling points for studying spatial distribution of soil respiration. Thus, the BME algorithm, coupled with soil temperature data, can not only improve the accuracy of soil respiration spatial interpolation but can also reduce the number of sampling points.
Directory of Open Access Journals (Sweden)
Lozano, L. J.
2005-10-01
Full Text Available This work focuses on general guidelines to be considered for application of least-squares routines and artificial neural networks (ANN in the estimation of metal distribution equilibria in liquid-liquid extraction process. The goal of the procedure in the statistical method is to find the values of the equilibrium constants (K_{j} for the reactions involved in the metal extraction which minimizes the differences between experimental distribution coefficient (D_{exp} and theoretical distribution coefficients according to the mechanism proposed (D_{theor}- In the first part of the article, results obtained with the most frequently routine reported in the bibliography are compared with those obtained using the algorithms previously discussed. In the second part, the main features of a single back-propagation neural network for the same purpose are discussed, and the results obtained are compared with those obtained with the classical methods.
El trabajo presenta las líneas generales a considerar para la estimación del equilibrio de distribución de metales en procesos de extracción líquido-líquido, según dos métodos: algoritmo clásico de mínimos cuadrados y redes neuronales artificiales. El objetivo del procedimiento, en el caso del método estadístico, es encontrar los valores de las constantes de equilibrio (K_{j} para las reacciones involucradas en la extracción del metal, que minimizan las diferencias entre el coeficiente de distribución experimental y el coeficiente de distribución teórico, de acuerdo al mecanismo propuesto. En la primera parte del artículo se comparan los resultados obtenidos a partir de los algoritmos usados más habitualmente en la bibliografía, con los datos obtenidos mediante el algoritmo previamente descrito. En la segunda parte, se presentan las características fundamentales para aplicar una red neuronal sencilla con algoritmo back-propagation y los
Link adaptation algorithm for distributed coded transmissions in cooperative OFDMA systems
DEFF Research Database (Denmark)
Varga, Mihaly; Badiu, Mihai Alin; Bota, Vasile
2015-01-01
This paper proposes a link adaptation algorithm for cooperative transmissions in the down-link connection of an OFDMA-based wireless system. The algorithm aims at maximizing the spectral efficiency of a relay-aided communication link, while satisfying the block error rate constraints at both...... adaptation algorithm has linear complexity with the number of available resource blocks, while still provides a very good performance, as shown by simulation results....
Distributed parameter estimation in unreliable sensor networks via broadcast gossip algorithms.
Wang, Huiwei; Liao, Xiaofeng; Wang, Zidong; Huang, Tingwen; Chen, Guo
2016-01-01
In this paper, we present an asynchronous algorithm to estimate the unknown parameter under an unreliable network which allows new sensors to join and old sensors to leave, and can tolerate link failures. Each sensor has access to partially informative measurements when it is awakened. In addition, the proposed algorithm can avoid the interference among messages and effectively reduce the accumulated measurement and quantization errors. Based on the theory of stochastic approximation, we prove that our proposed algorithm almost surely converges to the unknown parameter. Finally, we present a numerical example to assess the performance and the communication cost of the algorithm. Copyright © 2015 Elsevier Ltd. All rights reserved.
Performance Evaluation of New Joint EDF-RM Scheduling Algorithm for Real Time Distributed System
Directory of Open Access Journals (Sweden)
Rashmi Sharma
2014-01-01
Full Text Available In Real Time System, the achievement of deadline is the main target of every scheduling algorithm. Earliest Deadline First (EDF, Rate Monotonic (RM, and least Laxity First are some renowned algorithms that work well in their own context. As we know, there is a very common problem Domino's effect in EDF that is generated due to overloading condition (EDF is not working well in overloading situation. Similarly, performance of RM is degraded in underloading condition. We can say that both algorithms are complements of each other. Deadline missing in both events happens because of their utilization bounding strategy. Therefore, in this paper we are proposing a new scheduling algorithm that carries through the drawback of both existing algorithms. Joint EDF-RM scheduling algorithm is implemented in global scheduler that permits task migration mechanism in between processors in the system. In order to check the improved behavior of proposed algorithm we perform simulation. Results are achieved and evaluated in terms of Success Ratio (SR, Average CPU Utilization (ECU, Failure Ratio (FR, and Maximum Tardiness parameters. In the end, the results are compared with the existing (EDF, RM, and D_R_EDF algorithms. It has been shown that the proposed algorithm performs better during overloading condition as well in underloading condition.
Zhu, Gaofeng; Li, Xin; Ma, Jinzhu; Wang, Yunquan; Liu, Shaomin; Huang, Chunlin; Zhang, Kun; Hu, Xiaoli
2018-04-01
Sequential Monte Carlo (SMC) samplers have become increasing popular for estimating the posterior parameter distribution with the non-linear dependency structures and multiple modes often present in hydrological models. However, the explorative capabilities and efficiency of the sampler depends strongly on the efficiency in the move step of SMC sampler. In this paper we presented a new SMC sampler entitled the Particle Evolution Metropolis Sequential Monte Carlo (PEM-SMC) algorithm, which is well suited to handle unknown static parameters of hydrologic model. The PEM-SMC sampler is inspired by the works of Liang and Wong (2001) and operates by incorporating the strengths of the genetic algorithm, differential evolution algorithm and Metropolis-Hasting algorithm into the framework of SMC. We also prove that the sampler admits the target distribution to be a stationary distribution. Two case studies including a multi-dimensional bimodal normal distribution and a conceptual rainfall-runoff hydrologic model by only considering parameter uncertainty and simultaneously considering parameter and input uncertainty show that PEM-SMC sampler is generally superior to other popular SMC algorithms in handling the high dimensional problems. The study also indicated that it may be important to account for model structural uncertainty by using multiplier different hydrological models in the SMC framework in future study.
Energy Technology Data Exchange (ETDEWEB)
Lee, Youngrok [Iowa State Univ., Ames, IA (United States)
2013-05-15
Heterogeneity exists on a data set when samples from di erent classes are merged into the data set. Finite mixture models can be used to represent a survival time distribution on heterogeneous patient group by the proportions of each class and by the survival time distribution within each class as well. The heterogeneous data set cannot be explicitly decomposed to homogeneous subgroups unless all the samples are precisely labeled by their origin classes; such impossibility of decomposition is a barrier to overcome for estimating nite mixture models. The expectation-maximization (EM) algorithm has been used to obtain maximum likelihood estimates of nite mixture models by soft-decomposition of heterogeneous samples without labels for a subset or the entire set of data. In medical surveillance databases we can find partially labeled data, that is, while not completely unlabeled there is only imprecise information about class values. In this study we propose new EM algorithms that take advantages of using such partial labels, and thus incorporate more information than traditional EM algorithms. We particularly propose four variants of the EM algorithm named EM-OCML, EM-PCML, EM-HCML and EM-CPCML, each of which assumes a specific mechanism of missing class values. We conducted a simulation study on exponential survival trees with five classes and showed that the advantages of incorporating substantial amount of partially labeled data can be highly signi cant. We also showed model selection based on AIC values fairly works to select the best proposed algorithm on each specific data set. A case study on a real-world data set of gastric cancer provided by Surveillance, Epidemiology and End Results (SEER) program showed a superiority of EM-CPCML to not only the other proposed EM algorithms but also conventional supervised, unsupervised and semi-supervised learning algorithms.
Directory of Open Access Journals (Sweden)
Shuai Zeng
2013-01-01
Full Text Available With the development of wireless technologies, mobile communication applies more and more extensively in the various walks of life. The social network of both fixed and mobile users can be seen as networked agent system. At present, kinds of devices and access network technology are widely used. Different users in this networked agent system may need different coding rates multimedia data due to their heterogeneous demand. This paper proposes a distributed flow rate control algorithm to optimize multimedia data transmission of the networked agent system with the coexisting various coding rates. In this proposed algorithm, transmission path and upload bandwidth of different coding rate data between source node, fixed and mobile nodes are appropriately arranged and controlled. On the one hand, this algorithm can provide user nodes with differentiated coding rate data and corresponding flow rate. On the other hand, it makes the different coding rate data and user nodes networked, which realizes the sharing of upload bandwidth of user nodes which require different coding rate data. The study conducts mathematical modeling on the proposed algorithm and compares the system that adopts the proposed algorithm with the existing system based on the simulation experiment and mathematical analysis. The results show that the system that adopts the proposed algorithm achieves higher upload bandwidth utilization of user nodes and lower upload bandwidth consumption of source node.
The primary advantage of Dynamically Dimensioned Search algorithm (DDS) is that it outperforms many other optimization techniques in both convergence speed and the ability in searching for parameter sets that satisfy statistical guidelines while requiring only one algorithm parameter (perturbation f...
DEFF Research Database (Denmark)
Jakobsen, Mikkel Koefoed; Madsen, Jan; Hansen, Michael Reichhardt
2010-01-01
One of the key design goals in Wireless Sensor Networks is long lasting or even continuous operation. Continuous operation is made possible through energy harvesting. Keeping the network operational imposes a demand to prevent network segmentation and power loss in nodes. It is therefore important...... that the best energy-wise route is found for each data transfer from a source node to the sink node. We present a new adaptive and distributed routing algorithm for finding energy optimised routes in a wireless sensor network with energy harvesting. The algorithm finds an energy efficient route from each source...
EM algorithm for one-shot device testing with competing risks under exponential distribution
International Nuclear Information System (INIS)
Balakrishnan, N.; So, H.Y.; Ling, M.H.
2015-01-01
This paper provides an extension of the work of Balakrishnan and Ling [1] by introducing a competing risks model into a one-shot device testing analysis under an accelerated life test setting. An Expectation Maximization (EM) algorithm is then developed for the estimation of the model parameters. An extensive Monte Carlo simulation study is carried out to assess the performance of the EM algorithm and then compare the obtained results with the initial estimates obtained by the Inequality Constrained Least Squares (ICLS) method of estimation. Finally, we apply the EM algorithm to a clinical data, ED01, to illustrate the method of inference developed here. - Highlights: • ALT data analysis for one-shot devices with competing risks is considered. • EM algorithm is developed for the determination of the MLEs. • The estimations of lifetime under normal operating conditions are presented. • The EM algorithm improves the convergence rate
Efficient sampling algorithms for Monte Carlo based treatment planning
International Nuclear Information System (INIS)
DeMarco, J.J.; Solberg, T.D.; Chetty, I.; Smathers, J.B.
1998-01-01
Efficient sampling algorithms are necessary for producing a fast Monte Carlo based treatment planning code. This study evaluates several aspects of a photon-based tracking scheme and the effect of optimal sampling algorithms on the efficiency of the code. Four areas were tested: pseudo-random number generation, generalized sampling of a discrete distribution, sampling from the exponential distribution, and delta scattering as applied to photon transport through a heterogeneous simulation geometry. Generalized sampling of a discrete distribution using the cutpoint method can produce speedup gains of one order of magnitude versus conventional sequential sampling. Photon transport modifications based upon the delta scattering method were implemented and compared with a conventional boundary and collision checking algorithm. The delta scattering algorithm is faster by a factor of six versus the conventional algorithm for a boundary size of 5 mm within a heterogeneous geometry. A comparison of portable pseudo-random number algorithms and exponential sampling techniques is also discussed
Bandyopadhyay, Saptarshi
Multi-agent systems are widely used for constructing a desired formation shape, exploring an area, surveillance, coverage, and other cooperative tasks. This dissertation introduces novel algorithms in the three main areas of shape formation, distributed estimation, and attitude control of large-scale multi-agent systems. In the first part of this dissertation, we address the problem of shape formation for thousands to millions of agents. Here, we present two novel algorithms for guiding a large-scale swarm of robotic systems into a desired formation shape in a distributed and scalable manner. These probabilistic swarm guidance algorithms adopt an Eulerian framework, where the physical space is partitioned into bins and the swarm's density distribution over each bin is controlled using tunable Markov chains. In the first algorithm - Probabilistic Swarm Guidance using Inhomogeneous Markov Chains (PSG-IMC) - each agent determines its bin transition probabilities using a time-inhomogeneous Markov chain that is constructed in real-time using feedback from the current swarm distribution. This PSG-IMC algorithm minimizes the expected cost of the transitions required to achieve and maintain the desired formation shape, even when agents are added to or removed from the swarm. The algorithm scales well with a large number of agents and complex formation shapes, and can also be adapted for area exploration applications. In the second algorithm - Probabilistic Swarm Guidance using Optimal Transport (PSG-OT) - each agent determines its bin transition probabilities by solving an optimal transport problem, which is recast as a linear program. In the presence of perfect feedback of the current swarm distribution, this algorithm minimizes the given cost function, guarantees faster convergence, reduces the number of transitions for achieving the desired formation, and is robust to disturbances or damages to the formation. We demonstrate the effectiveness of these two proposed swarm
A Bayesian sequential design using alpha spending function to control type I error.
Zhu, Han; Yu, Qingzhao
2017-10-01
We propose in this article a Bayesian sequential design using alpha spending functions to control the overall type I error in phase III clinical trials. We provide algorithms to calculate critical values, power, and sample sizes for the proposed design. Sensitivity analysis is implemented to check the effects from different prior distributions, and conservative priors are recommended. We compare the power and actual sample sizes of the proposed Bayesian sequential design with different alpha spending functions through simulations. We also compare the power of the proposed method with frequentist sequential design using the same alpha spending function. Simulations show that, at the same sample size, the proposed method provides larger power than the corresponding frequentist sequential design. It also has larger power than traditional Bayesian sequential design which sets equal critical values for all interim analyses. When compared with other alpha spending functions, O'Brien-Fleming alpha spending function has the largest power and is the most conservative in terms that at the same sample size, the null hypothesis is the least likely to be rejected at early stage of clinical trials. And finally, we show that adding a step of stop for futility in the Bayesian sequential design can reduce the overall type I error and reduce the actual sample sizes.
Directory of Open Access Journals (Sweden)
John E. Candelo-Becerra
2015-07-01
Full Text Available Electric power losses have a significant impact on the total costs of distribution networks. The use of renewable energy sources is a major alternative to improve power losses and costs, although other important issues are also enhanced such as voltage magnitudes and network congestion. However, determining the best location and size of renewable energy generators can be sometimes a challenging task due to a large number of possible combinations in the search space. Furthermore, the multiobjective functions increase the complexity of the problem and metaheuristics are preferred to find solutions in a relatively short time. This paper evaluates the performance of the cuckoo search (CS, harmony search (HS, and bat-inspired (BA algorithms for the location and size of renewable distributed generation (RDG in radial distribution networks using a multiobjective function defined as minimizing the energy losses and the RDG costs. The metaheuristic algorithms were programmed in Matlab and tested using the 33-node radial distribution network. The three algorithms obtained similar results for the two objectives evaluated, finding points close to the best solutions in the Pareto front. Comparisons showed that the CS obtained the minimum results for most points evaluated, but the BA and the HS were close to the best solution.
Niu, Chun-Yang; Qi, Hong; Huang, Xing; Ruan, Li-Ming; Wang, Wei; Tan, He-Ping
2015-11-01
A hybrid least-square QR decomposition (LSQR)-particle swarm optimization (LSQR-PSO) algorithm was developed to estimate the three-dimensional (3D) temperature distributions and absorption coefficients simultaneously. The outgoing radiative intensities at the boundary surface of the absorbing media were simulated by the line-of-sight (LOS) method, which served as the input for the inverse analysis. The retrieval results showed that the 3D temperature distributions of the participating media with known radiative properties could be retrieved accurately using the LSQR algorithm, even with noisy data. For the participating media with unknown radiative properties, the 3D temperature distributions and absorption coefficients could be retrieved accurately using the LSQR-PSO algorithm even with measurement errors. It was also found that the temperature field could be estimated more accurately than the absorption coefficients. In order to gain insight into the effects on the accuracy of temperature distribution reconstruction, the selection of the detection direction and the angle between two detection directions was also analyzed. Project supported by the Major National Scientific Instruments and Equipment Development Special Foundation of China (Grant No. 51327803), the National Natural Science Foundation of China (Grant No. 51476043), and the Fund of Tianjin Key Laboratory of Civil Aircraft Airworthiness and Maintenance in Civil Aviation University of China.
Kim, Seonghoon
2013-01-01
With known item response theory (IRT) item parameters, Lord and Wingersky provided a recursive algorithm for computing the conditional frequency distribution of number-correct test scores, given proficiency. This article presents a generalized algorithm for computing the conditional distribution of summed test scores involving real-number item…
CSIR Research Space (South Africa)
Abu-Mahfouz, Adnan M
2013-05-01
Full Text Available Determining the location of nodes is a key part of wireless sensor networks (WSNs). Many WSN applications require knowledge of nodes’ locations to perform their functions successfully. Several localisation algorithms rely on using all or most...
Bulyha, Alena; Heitzinger, Clemens
2011-01-01
In this work, a Monte-Carlo algorithm in the constant-voltage ensemble for the calculation of 3d charge concentrations at charged surfaces functionalized with biomolecules is presented. The motivation for this work is the theoretical understanding
Wang, Bingjie; Pi, Shaohua; Sun, Qi; Jia, Bo
2015-05-01
An improved classification algorithm that considers multiscale wavelet packet Shannon entropy is proposed. Decomposition coefficients at all levels are obtained to build the initial Shannon entropy feature vector. After subtracting the Shannon entropy map of the background signal, components of the strongest discriminating power in the initial feature vector are picked out to rebuild the Shannon entropy feature vector, which is transferred to radial basis function (RBF) neural network for classification. Four types of man-made vibrational intrusion signals are recorded based on a modified Sagnac interferometer. The performance of the improved classification algorithm has been evaluated by the classification experiments via RBF neural network under different diffusion coefficients. An 85% classification accuracy rate is achieved, which is higher than the other common algorithms. The classification results show that this improved classification algorithm can be used to classify vibrational intrusion signals in an automatic real-time monitoring system.
Sequential Quadratic Programming Algorithms for Optimization
1989-08-01
quadratic program- ma ng (SQ(2l ) aIiatain.seenis to be relgarded aIs tie( buest choice for the solution of smiall. dlense problema (see S tour L)toS...For the step along d, note that a < nOing + 3 szH + i3.ninA A a K f~Iz,;nd and from Id1 _< ,,, we must have that for some /3 , np , 11P11 < dn"p. 5.2...Nevertheless, many of these problems are considered hard to solve. Moreover, for some of these problems the assumptions made in Chapter 2 to establish the
Directory of Open Access Journals (Sweden)
Takahashi Wataru
2012-02-01
Full Text Available Abstract Background The purpose of this study was to compare dose distributions from three different algorithms with the x-ray Voxel Monte Carlo (XVMC calculations, in actual computed tomography (CT scans for use in stereotactic radiotherapy (SRT of small lung cancers. Methods Slow CT scan of 20 patients was performed and the internal target volume (ITV was delineated on Pinnacle3. All plans were first calculated with a scatter homogeneous mode (SHM which is compatible with Clarkson algorithm using Pinnacle3 treatment planning system (TPS. The planned dose was 48 Gy in 4 fractions. In a second step, the CT images, structures and beam data were exported to other treatment planning systems (TPSs. Collapsed cone convolution (CCC from Pinnacle3, superposition (SP from XiO, and XVMC from Monaco were used for recalculating. The dose distributions and the Dose Volume Histograms (DVHs were compared with each other. Results The phantom test revealed that all algorithms could reproduce the measured data within 1% except for the SHM with inhomogeneous phantom. For the patient study, the SHM greatly overestimated the isocenter (IC doses and the minimal dose received by 95% of the PTV (PTV95 compared to XVMC. The differences in mean doses were 2.96 Gy (6.17% for IC and 5.02 Gy (11.18% for PTV95. The DVH's and dose distributions with CCC and SP were in agreement with those obtained by XVMC. The average differences in IC doses between CCC and XVMC, and SP and XVMC were -1.14% (p = 0.17, and -2.67% (p = 0.0036, respectively. Conclusions Our work clearly confirms that the actual practice of relying solely on a Clarkson algorithm may be inappropriate for SRT planning. Meanwhile, CCC and SP were close to XVMC simulations and actual dose distributions obtained in lung SRT.
A DISTRIBUTED ENERGY EFFICIENT CLUSTERING ALGORITHM FOR DATA AGGREGATION IN WIRELESS SENSOR NETWORKS
Directory of Open Access Journals (Sweden)
Seyed Mohammad Bagher Musavi Shirazi
2018-06-01
Full Text Available Wireless sensor networks (WSNs are a new generation of networks typically consisting of a large number of inexpensive nodes with wireless communications. The main purpose of these networks is to collect information from the environment for further processing. Nodes in the network have been equipped with limited battery lifetime, so energy saving is one of the major issues in WSNs. If we balance the load among cluster heads and prevent having an extra load on just a few nodes in the network, we can reach longer network lifetime. One solution to control energy consumption and balance the load among nodes is to use clustering techniques. In this paper, we propose a new distributed energy-efficient clustering algorithm for data aggregation in wireless sensor networks, called Distributed Clustering for Data Aggregation (DCDA. In our new approach, an optimal transmission tree is constructed among sensor nodes with a new greedy method. Base station (BS is the root, cluster heads (CHs and relay nodes are intermediate nodes, and other nodes (cluster member nodes are the leaves of this transmission tree. DCDA balances load among CHs in intra-cluster and inter-cluster data communications using different cluster sizes. For efficient inter-cluster communications, some relay nodes will transfer data between CHs. Energy consumption, distance to the base station, and cluster heads’ centric metric are three main adjustment parameters for the cluster heads election. Simulation results show that the proposed protocol leads to the reduction of individual sensor nodes’ energy consumption and prolongs network lifetime, in comparison with other known methods. ABSTRAK: Rangkaian sensor wayarles (WSN adalah rangkaian generasi baru yang terdiri daripada nod-nod murah komunikasi wayarles. Tujuan rangkaian-rangkaian ini adalah bagi mengumpul maklumat sekeliling untuk proses seterusnya. Nod dalam rangkaian ini dilengkapi bateri kurang jangka hayat, jadi simpanan tenaga
Directory of Open Access Journals (Sweden)
Zhenlai You
2014-01-01
Full Text Available This paper introduces the milk-run schema into the express distribution logistics through the feasibility analysis of application of cyclic goods-taking schema in the express industry. In order to reach fully loaded as far as possible in distribution, the article improves the traditional model, adopting multi-objective decision and conforming constraint conditions to Milk-run schema, even approximating the practical truth solves model through the C-W saving algorithm. It can effectively shorten the distance and Lower costs by means of reasonable route planning. Finally, the paper has verified the model and its effectiveness of arithmetic application by means of cases analysis.
Yang, Dixiong; Liu, Zhenjun; Zhou, Jilei
2014-04-01
Chaos optimization algorithms (COAs) usually utilize the chaotic map like Logistic map to generate the pseudo-random numbers mapped as the design variables for global optimization. Many existing researches indicated that COA can more easily escape from the local minima than classical stochastic optimization algorithms. This paper reveals the inherent mechanism of high efficiency and superior performance of COA, from a new perspective of both the probability distribution property and search speed of chaotic sequences generated by different chaotic maps. The statistical property and search speed of chaotic sequences are represented by the probability density function (PDF) and the Lyapunov exponent, respectively. Meanwhile, the computational performances of hybrid chaos-BFGS algorithms based on eight one-dimensional chaotic maps with different PDF and Lyapunov exponents are compared, in which BFGS is a quasi-Newton method for local optimization. Moreover, several multimodal benchmark examples illustrate that, the probability distribution property and search speed of chaotic sequences from different chaotic maps significantly affect the global searching capability and optimization efficiency of COA. To achieve the high efficiency of COA, it is recommended to adopt the appropriate chaotic map generating the desired chaotic sequences with uniform or nearly uniform probability distribution and large Lyapunov exponent.
International Nuclear Information System (INIS)
Ahmadigorji, Masoud; Amjady, Nima
2016-01-01
In this paper, a new model for MEPDN (multiyear expansion planning of distribution networks) is proposed. By solving this model, the optimal expansion scheme of primary (i.e. medium voltage) distribution network including the reinforcement pattern of primary feeders as well as location and size of DG (distributed generators) during an ascertained planning period is determined. Furthermore, the time-based feature of proposed model allows it to specify the investments/reinforcements time (i.e. year). Moreover, a minimum load shedding-based analytical approach for optimizing the network's reliability is introduced. The associated objective function of proposed model is minimizing the total investment and operation costs. To solve the formulated MEPDN model as a complex multi-dimensional optimization problem, a new evolutionary algorithm-based solution method called BCSSO (Binary Chaotic Shark Smell Optimization) is presented. The effectiveness of the proposed MEPDN model and solution approach is illustrated by applying them on two widely-used test cases including 12-bus and 33-bus distribution network and comparing the acquired results with the results of other solution methods. - Highlights: • A multiyear expansion planning model for distribution network is presented. • A new evolutionary algorithm-based solution approach is proposed. • A minimum load shedding-based analytical method for EENS minimization is suggested. • The efficacy of the proposed solution approach is broadly investigated.
On the effect of response transformations in sequential parameter optimization.
Wagner, Tobias; Wessing, Simon
2012-01-01
Parameter tuning of evolutionary algorithms (EAs) is attracting more and more interest. In particular, the sequential parameter optimization (SPO) framework for the model-assisted tuning of stochastic optimizers has resulted in established parameter tuning algorithms. In this paper, we enhance the SPO framework by introducing transformation steps before the response aggregation and before the actual modeling. Based on design-of-experiments techniques, we empirically analyze the effect of integrating different transformations. We show that in particular, a rank transformation of the responses provides significant improvements. A deeper analysis of the resulting models and additional experiments with adaptive procedures indicates that the rank and the Box-Cox transformation are able to improve the properties of the resultant distributions with respect to symmetry and normality of the residuals. Moreover, model-based effect plots document a higher discriminatory power obtained by the rank transformation.
Multi-agent sequential hypothesis testing
Kim, Kwang-Ki K.
2014-12-15
This paper considers multi-agent sequential hypothesis testing and presents a framework for strategic learning in sequential games with explicit consideration of both temporal and spatial coordination. The associated Bayes risk functions explicitly incorporate costs of taking private/public measurements, costs of time-difference and disagreement in actions of agents, and costs of false declaration/choices in the sequential hypothesis testing. The corresponding sequential decision processes have well-defined value functions with respect to (a) the belief states for the case of conditional independent private noisy measurements that are also assumed to be independent identically distributed over time, and (b) the information states for the case of correlated private noisy measurements. A sequential investment game of strategic coordination and delay is also discussed as an application of the proposed strategic learning rules.
Distributed topology control algorithm to conserve energy in heterogeneous wireless mesh networks
CSIR Research Space (South Africa)
Aron, FO
2008-07-01
Full Text Available in performance with the resulting topology being a sub-network of the one generated by [11]. Li and Halpern [13] further propose the small minimum energy communication network (SMECN). In this algorithm, each node u initially broadcasts a “hello” message... algorithm that runs in each node is presented as follows:- Phase1: Establishing the accessible neighbourhood topology. In this stage, node u broadcasts a “hello” message using its full power, max uP . The nodes that receive the “hello” message form...
Adaptive sequential controller
Energy Technology Data Exchange (ETDEWEB)
El-Sharkawi, Mohamed A. (Renton, WA); Xing, Jian (Seattle, WA); Butler, Nicholas G. (Newberg, OR); Rodriguez, Alonso (Pasadena, CA)
1994-01-01
An adaptive sequential controller (50/50') for controlling a circuit breaker (52) or other switching device to substantially eliminate transients on a distribution line caused by closing and opening the circuit breaker. The device adaptively compensates for changes in the response time of the circuit breaker due to aging and environmental effects. A potential transformer (70) provides a reference signal corresponding to the zero crossing of the voltage waveform, and a phase shift comparator circuit (96) compares the reference signal to the time at which any transient was produced when the circuit breaker closed, producing a signal indicative of the adaptive adjustment that should be made. Similarly, in controlling the opening of the circuit breaker, a current transformer (88) provides a reference signal that is compared against the time at which any transient is detected when the circuit breaker last opened. An adaptive adjustment circuit (102) produces a compensation time that is appropriately modified to account for changes in the circuit breaker response, including the effect of ambient conditions and aging. When next opened or closed, the circuit breaker is activated at an appropriately compensated time, so that it closes when the voltage crosses zero and opens when the current crosses zero, minimizing any transients on the distribution line. Phase angle can be used to control the opening of the circuit breaker relative to the reference signal provided by the potential transformer.
Adaptive sequential controller
El-Sharkawi, Mohamed A.; Xing, Jian; Butler, Nicholas G.; Rodriguez, Alonso
1994-01-01
An adaptive sequential controller (50/50') for controlling a circuit breaker (52) or other switching device to substantially eliminate transients on a distribution line caused by closing and opening the circuit breaker. The device adaptively compensates for changes in the response time of the circuit breaker due to aging and environmental effects. A potential transformer (70) provides a reference signal corresponding to the zero crossing of the voltage waveform, and a phase shift comparator circuit (96) compares the reference signal to the time at which any transient was produced when the circuit breaker closed, producing a signal indicative of the adaptive adjustment that should be made. Similarly, in controlling the opening of the circuit breaker, a current transformer (88) provides a reference signal that is compared against the time at which any transient is detected when the circuit breaker last opened. An adaptive adjustment circuit (102) produces a compensation time that is appropriately modified to account for changes in the circuit breaker response, including the effect of ambient conditions and aging. When next opened or closed, the circuit breaker is activated at an appropriately compensated time, so that it closes when the voltage crosses zero and opens when the current crosses zero, minimizing any transients on the distribution line. Phase angle can be used to control the opening of the circuit breaker relative to the reference signal provided by the potential transformer.
Optimal Power Flow in Islanded Microgrids Using a Simple Distributed Algorithm
DEFF Research Database (Denmark)
Sanseverino, Eleonora Riva; Di Silvestre, Maria Luisa; Badalamenti, Romina
2015-01-01
In this paper, the problem of distributed power losses minimization in islanded distribution systems is dealt with. The problem is formulated in a very simple manner and a solution is reached after a few iterations. The considered distribution system, a microgrid, will not need large bandwidth co...
Degree distribution of shortest path trees and bias of network sampling algorithms
Bhamidi, S.; Goodman, J.A.; Hofstad, van der R.W.; Komjáthy, J.
2013-01-01
In this article, we explicitly derive the limiting distribution of the degree distribution of the shortest path tree from a single source on various random network models with edge weights. We determine the power-law exponent of the degree distribution of this tree and compare it to the degree
Degree distribution of shortest path trees and bias of network sampling algorithms
Bhamidi, S.; Goodman, J.A.; Hofstad, van der R.W.; Komjáthy, J.
2015-01-01
In this article, we explicitly derive the limiting degree distribution of the shortest path tree from a single source on various random network models with edge weights. We determine the asymptotics of the degree distribution for large degrees of this tree and compare it to the degree distribution
CSIR Research Space (South Africa)
Machaka, Pheeha
2015-11-01
Full Text Available The paper presents the threats that are present in Internet of Things (IoT) systems and how they can be used to perpetuate a large scale DDoS attack. The paper investigates how the Cumulative Sum (CUSUM) algorithm can be used to detect a DDoS attack...
Directory of Open Access Journals (Sweden)
Tamilselvan V.
2016-06-01
Full Text Available The radial distribution system is a rugged system, it is also the most commonly used system, which suffers by loss and low voltage at the end bus. This loss can be reduced by the use of a capacitor in the system, which injects reactive current and also improves the voltage magnitude in the buses. The real power loss in the distribution line is the I2R loss which depends on the current and resistance. The connection of the capacitor in the bus reduces the reactive current and losses. The loss reduction is equal to the increase in generation, necessary for the electric power provided by firms. For consumers, the quality of power supply depends on the voltage magnitude level, which is also considered and hence the objective of the problem becomes the multi objective of loss minimization and the minimization of voltage deviation. In this paper, the optimal location and size of the capacitor is found using a new computational intelligent algorithm called Flower Pollination Algorithm (FPA. To calculate the power flow and losses in the system, novel data structure load flow is introduced. In this, each bus is considered as a node with bus associated data. Links between the nodes are distribution lines and their own resistance and reactance. To validate the developed FPA solutions standard test cases, IEEE 33 and IEEE 69 radial distribution systems are considered.
Radhakrishnan, Mohanasundar; Pathirana, Assela; Ghebremichael, Kebreab A.; Amy, Gary L.
2012-01-01
Concerns have been raised regarding disinfection by-products (DBPs) formed as a result of the reaction of halogen-based disinfectants with DBP precursors. In order to appreciate the chemical and biological tradeoffs, it is imperative to understand the formation trends of DBPs and their spread in the distribution network. However, the water at a point in a complex distribution system is a mixture from various sources, whose proportions are complex to estimate and requires advanced hydraulic analysis. To understand the risks of DBPs and to develop mitigation strategies, it is important to understand the distribution of DBPs in a water network, which requires modelling. The goal of this research was to integrate a steady-state water network model with a particle backtracking algorithm and chlorination as well as DBPs models in order to assess the tradeoffs between biological and chemical risks in the distribution network. A multi-objective optimisation algorithm was used to identify the optimal proportion of water from various sources, dosages of alum, and dosages of chlorine in the treatment plant and in booster locations to control the formation of chlorination DBPs and to achieve a balance between microbial and chemical risks. © IWA Publishing 2012.
Radhakrishnan, Mohanasundar
2012-05-01
Concerns have been raised regarding disinfection by-products (DBPs) formed as a result of the reaction of halogen-based disinfectants with DBP precursors. In order to appreciate the chemical and biological tradeoffs, it is imperative to understand the formation trends of DBPs and their spread in the distribution network. However, the water at a point in a complex distribution system is a mixture from various sources, whose proportions are complex to estimate and requires advanced hydraulic analysis. To understand the risks of DBPs and to develop mitigation strategies, it is important to understand the distribution of DBPs in a water network, which requires modelling. The goal of this research was to integrate a steady-state water network model with a particle backtracking algorithm and chlorination as well as DBPs models in order to assess the tradeoffs between biological and chemical risks in the distribution network. A multi-objective optimisation algorithm was used to identify the optimal proportion of water from various sources, dosages of alum, and dosages of chlorine in the treatment plant and in booster locations to control the formation of chlorination DBPs and to achieve a balance between microbial and chemical risks. © IWA Publishing 2012.
Directory of Open Access Journals (Sweden)
Qi Hong
2015-01-01
Full Text Available The particle size distribution (PSD plays an important role in environmental pollution detection and human health protection, such as fog, haze and soot. In this study, the Attractive and Repulsive Particle Swarm Optimization (ARPSO algorithm and the basic PSO were applied to retrieve the PSD. The spectral extinction technique coupled with the Anomalous Diffraction Approximation (ADA and the Lambert-Beer Law were employed to investigate the retrieval of the PSD. Three commonly used monomodal PSDs, i.e. the Rosin-Rammer (R-R distribution, the normal (N-N distribution, the logarithmic normal (L-N distribution were studied in the dependent model. Then, an optimal wavelengths selection algorithm was proposed. To study the accuracy and robustness of the inverse results, some characteristic parameters were employed. The research revealed that the ARPSO showed more accurate and faster convergence rate than the basic PSO, even with random measurement error. Moreover, the investigation also demonstrated that the inverse results of four incident laser wavelengths showed more accurate and robust than those of two wavelengths. The research also found that if increasing the interval of the selected incident laser wavelengths, inverse results would show more accurate, even in the presence of random error.
Sequential decoders for large MIMO systems
Ali, Konpal S.; Abediseid, Walid; Alouini, Mohamed-Slim
2014-01-01
the Sequential Decoder using the Fano Algorithm for large MIMO systems. A parameter called the bias is varied to attain different performance-complexity trade-offs. Low values of the bias result in excellent performance but at the expense of high complexity
STABILIZED SEQUENTIAL QUADRATIC PROGRAMMING: A SURVEY
Directory of Open Access Journals (Sweden)
Damián Fernández
2014-12-01
Full Text Available We review the motivation for, the current state-of-the-art in convergence results, and some open questions concerning the stabilized version of the sequential quadratic programming algorithm for constrained optimization. We also discuss the tools required for its local convergence analysis, globalization challenges, and extentions of the method to the more general variational problems.
Zips : mining compressing sequential patterns in streams
Hoang, T.L.; Calders, T.G.K.; Yang, J.; Mörchen, F.; Fradkin, D.; Chau, D.H.; Vreeken, J.; Leeuwen, van M.; Faloutsos, C.
2013-01-01
We propose a streaming algorithm, based on the minimal description length (MDL) principle, for extracting non-redundant sequential patterns. For static databases, the MDL-based approach that selects patterns based on their capacity to compress data rather than their frequency, was shown to be
DEFF Research Database (Denmark)
Cordua, Knud Skou; Hansen, Thomas Mejer; Lange, Katrine
In order to move beyond simplified covariance based a priori models, which are typically used for inverse problems, more complex multiple-point-based a priori models have to be considered. By means of marginal probability distributions ‘learned’ from a training image, sequential simulation has...... proven to be an efficient way of obtaining multiple realizations that honor the same multiple-point statistics as the training image. The frequency matching method provides an alternative way of formulating multiple-point-based a priori models. In this strategy the pattern frequency distributions (i.......e. marginals) of the training image and a subsurface model are matched in order to obtain a solution with the same multiple-point statistics as the training image. Sequential Gibbs sampling is a simulation strategy that provides an efficient way of applying sequential simulation based algorithms as a priori...
International Nuclear Information System (INIS)
Liu, Lei; Liu, Shuntao; Yuan, Fuh-Gwo
2012-01-01
A distributed on-board algorithm that is embedded and executed within a group of wireless sensors to locate structural damages in isotropic plates is presented. The algorithm is based on an energy-decay model of Lamb waves and singular value decomposition (SVD) to determine damage locations. A sensor group consists of a small number of sensors, each of which independently collects wave signals and evaluates wave energy upon an external triggering signal sent from a base station. The energy values, usually a few bytes in length, are then sent to the base station to determine the presence and location of damages. In comparison with traditional centralized approaches in which whole datasets are required to be transmitted, the proposed algorithm yields much less wireless communication traffic, yet with a modest amount of computation required within sensors. Experiments have shown that the algorithm is robust to locate damage for isotropic plate structures and is very power efficient, with more than an order-of-magnitude power saving
Xiao, Chuan-Le; Chen, Xiao-Zhou; Du, Yang-Li; Sun, Xuesong; Zhang, Gong; He, Qing-Yu
2013-01-04
Mass spectrometry has become one of the most important technologies in proteomic analysis. Tandem mass spectrometry (LC-MS/MS) is a major tool for the analysis of peptide mixtures from protein samples. The key step of MS data processing is the identification of peptides from experimental spectra by searching public sequence databases. Although a number of algorithms to identify peptides from MS/MS data have been already proposed, e.g. Sequest, OMSSA, X!Tandem, Mascot, etc., they are mainly based on statistical models considering only peak-matches between experimental and theoretical spectra, but not peak intensity information. Moreover, different algorithms gave different results from the same MS data, implying their probable incompleteness and questionable reproducibility. We developed a novel peptide identification algorithm, ProVerB, based on a binomial probability distribution model of protein tandem mass spectrometry combined with a new scoring function, making full use of peak intensity information and, thus, enhancing the ability of identification. Compared with Mascot, Sequest, and SQID, ProVerB identified significantly more peptides from LC-MS/MS data sets than the current algorithms at 1% False Discovery Rate (FDR) and provided more confident peptide identifications. ProVerB is also compatible with various platforms and experimental data sets, showing its robustness and versatility. The open-source program ProVerB is available at http://bioinformatics.jnu.edu.cn/software/proverb/ .
International Nuclear Information System (INIS)
Shivaie, Mojtaba; Ameli, Mohammad T.; Sepasian, Mohammad S.; Weinsier, Philip D.; Vahidinasab, Vahid
2015-01-01
In this paper, the authors present a new multistage framework for reliability-based Distribution Expansion Planning (DEP) in which expansion options are a reinforcement and/or installation of substations, feeders, and Distributed Generations (DGs). The proposed framework takes into account not only costs associated with investment, maintenance, and operation, but also expected customer interruption cost in the optimization as four problem objectives. At the same time, operational restrictions, Kirchhoff's laws, radial structure limitation, voltage limits, and capital expenditure budget restriction are considered as problem constraints. The proposed model is a non-convex optimization problem having a non-linear, mixed-integer nature. Hence, a hybrid Self-adaptive Global-based Harmony Search Algorithm (SGHSA) and Optimal Power Flow (OPF) were used and followed by a fuzzy satisfying method in order to obtain the final optimal solution. The SGHSA is a recently developed optimization algorithm which imitates the music improvisation process. In this process, the harmonists improvise their instrument pitches, searching for the perfect state of harmony. The planning methodology was demonstrated on the 27-node, 13.8-kV test system in order to demonstrate the feasibility and capability of the proposed model. Simulation results illustrated the sufficiency and profitableness of the newly developed framework, when compared with other methods. - Highlights: • A new multistage framework is presented for reliability-based DEP problem. • In this paper, DGs are considered as an expansion option to increase the flexibility of the proposed model. • In this paper, effective factors of DEP problem are incorporated as a multi-objective model. • In this paper, three new algorithms HSA, IHSA and SGHSA are proposed. • Results obtained by the proposed SGHSA algorithm are better than others
Energy Technology Data Exchange (ETDEWEB)
Parsakhoo, P.
2016-07-01
Aim of study: Corrected Backmund and Surface Distribution Algorithms (SDA) for analysis of forest road network are introduced and presented in this study. Research was carried out to compare road network performance between two districts in a hardwood forest. Area of study: Shast Kalateh forests, Iran. Materials and methods: In uncorrected Backmund algorithm, skidding distance was determined by calculating road density and spacing and then it was designed as Potential Area for Skidding Operations (PASO) in ArcGIS software. To correct this procedure, the skidding constraint areas were taken using GPS and then removed from PASO. In SDA, shortest perpendicular distance from geometrical center of timber compartments to road was measured at both districts. Main results: In corrected Backmund, forest openness in district I and II were 70.3% and 69.5%, respectively. Therefore, there was little difference in forest openness in the districts based on the uncorrected Backmund. In SDA, the mean distance from geometrical center of timber compartments to the roads of districts I and II were 199.45 and 149.31 meters, respectively. Forest road network distribution in district II was better than that of district I relating to SDA. Research highlights: It was concluded that uncorrected Backmund was not precise enough to assess forest road network, while corrected Backmund could exhibit a real PASO by removing skidding constraints. According to presented algorithms, forest road network performance in district II was better than district I. (Author)
Directory of Open Access Journals (Sweden)
Ru-Min Chao
2018-06-01
Full Text Available Due to the shortage of fossil fuel and the environmental pollution problem, solar energy applications have drawn a lot of attention worldwide. This paper reports the use of the latest patented distributed photovoltaic (PV power system design, including the two possible maximum power point tracking (MPPT algorithms, a power optimizer, and a PV power controller, in grid-connected and standalone applications. A distributed PV system with four amorphous silicon thin-film solar panels is used to evaluate both the quadratic maximization (QM and the Steepest descent (SD MPPT algorithms. The system’s design is different for the QM or the SD MPPT algorithm being used. The test result for the grid-connected silicon-based PV panels will also be reported. Considering the settling time for the power optimizer to be 20 ms, the test result shows that the tracking time for the QM method is close to 200 ms, which is faster when compared with the SD method whose tracking time is 500 ms. Besides this, the use of the QM method provides a more stable power output since the tracking is restricted by a local power optimizer rather than the global tracking the SD method uses. For a standalone PV application, a solar-powered boat design with 18 PV panels using a cascaded MPPT controller is introduced, and it provides flexibility in system design and the effective use of photovoltaic energy.
A Survey of Multi-Objective Sequential Decision-Making
Roijers, D.M.; Vamplew, P.; Whiteson, S.; Dazeley, R.
2013-01-01
Sequential decision-making problems with multiple objectives arise naturally in practice and pose unique challenges for research in decision-theoretic planning and learning, which has largely focused on single-objective settings. This article surveys algorithms designed for sequential
Energy Technology Data Exchange (ETDEWEB)
Stromberger, Carmen; Ghadjar, Pirus; Marnitz, Simone; Thieme, Alexander Henry; Jahn, Ulrich; Karaj-Rossbacher, Evis; Budach, Volker [Charite Universitaetsmedizin Berlin, Department of Radiation Oncology and Radiotherapy, Berlin (Germany); Raguse, Jan D. [Charite Universitaetsmedizin Berlin, Clinic for Oral and Maxillofacial Surgery, Berlin (Germany); Boettcher, Arne [Charite Universitaetsmedizin Berlin, Otorhinolaryngology, Berlin (Germany); Jamil, Basil [Communal Hospital Frankfurt Oder, Department of Radiation Oncology, Frankfurt/Oder (Germany)
2016-01-15
The purpose of this work was to compare sequential (SeqB) versus simultaneous integrated boost (SIB) radiotherapy plans delivered with volumetric modulated arc therapy (VMAT) for patients with locally advanced squamous cell cancer of the head and neck (HNSCC). SeqB and SIB plans using VMAT for 10 HNSCC patients given definitive chemoradiation were generated and analysed for differences in dose distribution, coverage, conformity and homogeneity to the planning target volumes (PTV) 1-3 and sparing of organs at risk (OAR). The mean delineated volumes ± standard deviations were 137.7 ± 44.8, 351.3 ± 83.9 and 895.6 ± 120.5 cm{sup 3} for PTV1-3. The mean volumes encompassed by the corresponding 95 % isodoses were 281 (+ 110 %) ± 73.4, 712.2 (+ 115 %) ± 146.4 and 1381.1 (+ 54 %) ± 217.3 cm{sup 3} with SeqB and 138.2 (+ 7 %) ± 40.1, 380.4 (+ 11 %) ± 91.9 and 1057.3 (+ 21 %) ± 161.4 cm{sup 3} with SIB for PTV1-3, respectively. Both strategies achieved excellent PTV coverage. SeqB provided significantly better coverage of PTV1 and 3, worse conformity for PTV1-3 and a higher mean dose than prescribed (111-115 %) to PTV2 and 3 (p ≤ 0.007). Both strategies provided satisfactory OAR sparing. This study showed significant dosimetric differences with potential clinical relevance between two VMAT boost strategies regarding coverage, conformity and dose to the PTVs. SIB might cause less toxicity. A clinical phase III/IV trial endorsed by the German Head and Neck Clinical Trials Group (IAG-KHT) will evaluate differences in acute/late toxicity as well as in locoregional recurrences between the two boost techniques. (orig.) [German] Vergleich von sequentiellem (SeqB) und simultan-integriertem Boost (SIB) mit moderner volumetrischer Arc-Therapie (VMAT) fuer Patienten mit Plattenepithelkarzinomen der Kopf-Hals-Region. Fuer 10 Patienten mit Plattenepithelkarzinomen der Kopf-Hals-Region und definitiver Radiochemotherapie erfolgte eine VMAT-Planung als SeqB und SIB fuer die
Directory of Open Access Journals (Sweden)
Zineb Simeu-Abazi
2009-11-01
Full Text Available The optimal management of a potable water distribution system requires the control of the reference (standard data, the control points, control of the drainage parameters (pressure, flow, etc. and maintenance parameters. The control of the mentioned data defines the network learning process [1]. Besides classic IT functions of acquisition, storage and data processing, a geographical information system (GIS can be used as the basis for an alarm system, allowing one to identify and to localize the presence of water leaks in the network [2]. In this article we propose an algorithm coupling the various drainage parameters for the management of the network. The algorithm leads to an optimal management of leaks. An application is in progress on the primary network in the region of Bonaberi in Douala, the largest city of Cameroon.
Xiang, Min; Qu, Qinqin; Chen, Cheng; Tian, Li; Zeng, Lingkang
2017-11-01
To improve the reliability of communication service in smart distribution grid (SDG), an access selection algorithm based on dynamic network status and different service types for heterogeneous wireless networks was proposed. The network performance index values were obtained in real time by multimode terminal and the variation trend of index values was analyzed by the growth matrix. The index weights were calculated by entropy-weight and then modified by rough set to get the final weights. Combining the grey relational analysis to sort the candidate networks, and the optimum communication network is selected. Simulation results show that the proposed algorithm can implement dynamically access selection in heterogeneous wireless networks of SDG effectively and reduce the network blocking probability.
DEFF Research Database (Denmark)
Meng, Lexuan; Zhao, Xin; Tang, Fen
2016-01-01
In islanded microgrids (MGs), distributed generators (DGs) can be employed as distributed compensators for improving the power quality in the consumer side. Two-level hierarchical control can be used for voltage unbalance compensation. Primary level, consisting of droop control and virtual...
DEFF Research Database (Denmark)
Meng, Lexuan; Dragicevic, Tomislav; Guerrero, Josep M.
2014-01-01
In a DC microgrid, several paralleled conversion systems are installed in distributed substations for transferring power from external grid to a DC microgrid. Droop control is used for the distributed load sharing among all the DC/DC converters. Considering the typical efficiency feature of power...
International Nuclear Information System (INIS)
Yamamoto, Akio; Hashimoto, Hiroshi
2002-01-01
The distributed genetic algorithm (DGA) is applied for loading pattern optimization problems of the pressurized water reactors. A basic concept of DGA follows that of the conventional genetic algorithm (GA). However, DGA equally distributes candidates of solutions (i.e. loading patterns) to several independent ''islands'' and evolves them in each island. Communications between islands, i.e. migrations of some candidates between islands are performed with a certain period. Since candidates of solutions independently evolve in each island while accepting different genes of migrants, premature convergence in the conventional GA can be prevented. Because many candidate loading patterns should be evaluated in GA or DGA, the parallelization is efficient to reduce turn around time. Parallel efficiency of DGA was measured using our optimization code and good efficiency was attained even in a heterogeneous cluster environment due to dynamic distribution of the calculation load. The optimization code is based on the client/server architecture with the TCP/IP native socket and a client (optimization) module and calculation server modules communicate the objects of loading patterns each other. Throughout the sensitivity study on optimization parameters of DGA, a suitable set of the parameters for a test problem was identified. Finally, optimization capability of DGA and the conventional GA was compared in the test problem and DGA provided better optimization results than the conventional GA. (author)
International Nuclear Information System (INIS)
Shao, Wei; Cui, Zheng; Cheng, Lin
2016-01-01
Highlights: • A multi-objective optimization model of air distribution of grate cooler by genetic algorithm is proposed. • Pareto Front is obtained and validated by comparing with operating data. • Optimal schemes are compared and selected by engineering background. • Total power consumption after optimization decreases 61.10%. • Thickness of clinker on three grate plates is thinner. - Abstract: The cooling air distributions of grate cooler exercise a great influence on the clinker cooling efficiency and power consumption of cooling fans. A multi-objective optimization model of air distributions of grate cooler with cross-flow heat exchanger analogy is proposed in this paper. Firstly, thermodynamic and flow models of clinker cooling process is carried out. Then based on entropy generation minimization analysis, modified entropy generation numbers caused by heat transfer and pressure drop are chosen as objective functions respectively which optimized by genetic algorithm. The design variables are superficial velocities of air chambers and thicknesses of clinker layers on different grate plates. A set of Pareto optimal solutions which two objectives are optimized simultaneously is achieved. Scattered distributions of design variables resulting in the conflict between two objectives are brought out. The final optimal air distribution and thicknesses of clinker layers are selected from the Pareto optimal solutions based on power consumption of cooling fans minimization and validated by measurements. Compared with actual operating scheme, the total air volumes of optimized schemes decrease 2.4%, total power consumption of cooling fans decreases 61.1% and the outlet temperature of clinker decreases 122.9 °C which shows a remarkable energy-saving effect on energy consumption.
DEFF Research Database (Denmark)
Aldana, Nelson Leonardo Diaz; Hernández, Adriana Carolina Luna; Vasquez, Juan Carlos
2015-01-01
This paper presents a centralized strategy for equalizing the state of charge of distributed energy storage systems in an islanded DC microgrid. The proposed strategy is based on a simple algorithm called equalization algorithm, which modifies the charge or discharge rate by weighting the virtual...... results of a low voltage DC microgrid are presented in order to verify the performance of the proposed approach....
Zhang, Y.; Chatterjea, Supriyo; Havinga, Paul J.M.
2007-01-01
We report our experiences with implementing a distributed and self-organizing scheduling algorithm designed for energy-efficient data gathering on a 25-node multihop wireless sensor network (WSN). The algorithm takes advantage of spatial correlations that exist in readings of adjacent sensor nodes
Construction and Application of an AMR Algorithm for Distributed Memory Computers
Deiterding, Ralf
2003-01-01
While the parallelization of blockstructured adaptive mesh refinement techniques is relatively straight-forward on shared memory architectures, appropriate distribution strategies for the emerging generation of distributed memory machines are a topic of on-going research. In this paper, a locality-preserving domain decomposition is proposed that partitions the entire AMR hierarchy from the base level on. It is shown that the approach reduces the communication costs and simplifies the im...
1983-10-01
an Aborti , It forwards the operation directly to the recovery system. When the recovery system acknowledges that the operation has been processed, the...list... AbortI . rite Ti Into the abort list. Then undo all of Ti’s writes by reedina their bet ore-images from the audit trail and writin. them back...Into the stable database. [Ack) Then, delete Ti from the active list. Restart. Process Aborti for each Ti on the active list. Ack) In this algorithm
Hortos, William S.
2008-04-01
Proposed distributed wavelet-based algorithms are a means to compress sensor data received at the nodes forming a wireless sensor network (WSN) by exchanging information between neighboring sensor nodes. Local collaboration among nodes compacts the measurements, yielding a reduced fused set with equivalent information at far fewer nodes. Nodes may be equipped with multiple sensor types, each capable of sensing distinct phenomena: thermal, humidity, chemical, voltage, or image signals with low or no frequency content as well as audio, seismic or video signals within defined frequency ranges. Compression of the multi-source data through wavelet-based methods, distributed at active nodes, reduces downstream processing and storage requirements along the paths to sink nodes; it also enables noise suppression and more energy-efficient query routing within the WSN. Targets are first detected by the multiple sensors; then wavelet compression and data fusion are applied to the target returns, followed by feature extraction from the reduced data; feature data are input to target recognition/classification routines; targets are tracked during their sojourns through the area monitored by the WSN. Algorithms to perform these tasks are implemented in a distributed manner, based on a partition of the WSN into clusters of nodes. In this work, a scheme of collaborative processing is applied for hierarchical data aggregation and decorrelation, based on the sensor data itself and any redundant information, enabled by a distributed, in-cluster wavelet transform with lifting that allows multiple levels of resolution. The wavelet-based compression algorithm significantly decreases RF bandwidth and other resource use in target processing tasks. Following wavelet compression, features are extracted. The objective of feature extraction is to maximize the probabilities of correct target classification based on multi-source sensor measurements, while minimizing the resource expenditures at
Effect of distributed generation installation on power loss using genetic algorithm method
Hasibuan, A.; Masri, S.; Othman, W. A. F. W. B.
2018-02-01
Injection of the generator distributed in the distribution network can affect the power system significantly. The effect that occurs depends on the allocation of DG on each part of the distribution network. Implementation of this approach has been made to the IEEE 30 bus standard and shows the optimum location and size of the DG which shows a decrease in power losses in the system. This paper aims to show the impact of distributed generation on the distribution system losses. The main purpose of installing DG on a distribution system is to reduce power losses on the power system.Some problems in power systems that can be solved with the installation of DG, one of which will be explored in the use of DG in this study is to reduce the power loss in the transmission line. Simulation results from case studies on the IEEE 30 bus standard system show that the system power loss decreased from 5.7781 MW to 1,5757 MW or just 27,27%. The simulated DG is injected to the bus with the lowest voltage drop on the bus number 8.
Izadi, Arman; Kimiagari, Ali Mohammad
2014-05-01
Distribution network design as a strategic decision has long-term effect on tactical and operational supply chain management. In this research, the location-allocation problem is studied under demand uncertainty. The purposes of this study were to specify the optimal number and location of distribution centers and to determine the allocation of customer demands to distribution centers. The main feature of this research is solving the model with unknown demand function which is suitable with the real-world problems. To consider the uncertainty, a set of possible scenarios for customer demands is created based on the Monte Carlo simulation. The coefficient of variation of costs is mentioned as a measure of risk and the most stable structure for firm's distribution network is defined based on the concept of robust optimization. The best structure is identified using genetic algorithms and 14 % reduction in total supply chain costs is the outcome. Moreover, it imposes the least cost variation created by fluctuation in customer demands (such as epidemic diseases outbreak in some areas of the country) to the logistical system. It is noteworthy that this research is done in one of the largest pharmaceutical distribution firms in Iran.
Distributed, signal strength-based indoor localization algorithm for use in healthcare environments.
Wyffels, Jeroen; De Brabanter, Jos; Crombez, Pieter; Verhoeve, Piet; Nauwelaers, Bart; De Strycker, Lieven
2014-11-01
In current healthcare environments, a trend toward mobile and personalized interactions between people and nurse call systems is strongly noticeable. Therefore, it should be possible to locate patients at all times and in all places throughout the care facility. This paper aims at describing a method by which a mobile node can locate itself indoors, based on signal strength measurements and a minimal amount of yes/no decisions. The algorithm has been developed specifically for use in a healthcare environment. With extensive testing and statistical support, we prove that our algorithm can be used in a healthcare setting with an envisioned level of localization accuracy up to room revel (or region level in a corridor), while avoiding heavy investments since the hardware of an existing nurse call network can be reused. The approach opted for leads to very high scalability, since thousands of mobile nodes can locate themselves. Network timing issues and localization update delays are avoided, which ensures that a patient can receive the needed care in a time and resources efficient way.
Direct numerical simulation of bubbles with adaptive mesh refinement with distributed algorithms
International Nuclear Information System (INIS)
Talpaert, Arthur
2017-01-01
This PhD work presents the implementation of the simulation of two-phase flows in conditions of water-cooled nuclear reactors, at the scale of individual bubbles. To achieve that, we study several models for Thermal-Hydraulic flows and we focus on a technique for the capture of the thin interface between liquid and vapour phases. We thus review some possible techniques for adaptive Mesh Refinement (AMR) and provide algorithmic and computational tools adapted to patch-based AMR, which aim is to locally improve the precision in regions of interest. More precisely, we introduce a patch-covering algorithm designed with balanced parallel computing in mind. This approach lets us finely capture changes located at the interface, as we show for advection test cases as well as for models with hyperbolic-elliptic coupling. The computations we present also include the simulation of the incompressible Navier-Stokes system, which models the shape changes of the interface between two non-miscible fluids. (author) [fr
International Nuclear Information System (INIS)
Tanabe, Akira; Yamamoto, Toru; Shinfuku, Kimihiro; Nakamae, Takuji; Nishide, Fusayo.
1995-01-01
Previously a two-layered neural network model was developed to predict the relation between fissile enrichment of each fuel rod and local power distribution in a BWR fuel bundle. This model was obtained intuitively based on 33 patterns of training signals after an intensive survey of the models. Recently, a learning algorithm with forgetting was reported to simplify neural network models. It is an interesting subject what kind of model will be obtained if this algorithm is applied to the complex three-layered model which learns the same training signals. A three-layered model which is expanded to have direct connections between the 1st and the 3rd layer elements has been constructed and the learning method of normal back propagation was applied first to this model. The forgetting algorithm was then added to this learning process. The connections concerned with the 2nd layer elements disappeared and the 2nd layer has become unnecessary. It took a longer computing time by an order to learn the same training signals than the simple back propagation, but the two-layered model was obtained autonomously from the expanded three-layered model. (author)
Amin, Nor Azrita Mohd; Adam, Mohd Bakri; Ibrahim, Noor Akma
2014-09-01
The Multiple-try Metropolis (MTM) algorithm is the new alternatives in the field of Bayesian extremes for summarizing the posterior distribution. MTM produce efficient estimation scheme for modelling extreme data in term of the convergence and small burn-in periods. The main objective is to explore the accuracy of the parameter estimation to the change of priors and compare the results with a classical likelihood-based analysis. Focus is on modelling the extreme data based on block maxima approach using Gumbel distribution. The comparative study between MTM and MLE is shown by the numerical problems. Several goodness of fit tests are compute for selecting the best model. The application is on the monthly maxima PM10 data for Johor state.
DEFF Research Database (Denmark)
Husfeldt, Thore
2015-01-01
This chapter presents an introduction to graph colouring algorithms. The focus is on vertex-colouring algorithms that work for general classes of graphs with worst-case performance guarantees in a sequential model of computation. The presentation aims to demonstrate the breadth of available...
Mannino, Antonio; Russ, Mary E.; Hooker, Stanford B.
2007-01-01
In coastal ocean waters, distributions of dissolved organic carbon (DOC) and chromophoric dissolved organic matter (CDOM) vary seasonally and interannually due to multiple source inputs and removal processes. We conducted several oceanographic cruises within the continental margin of the U.S. Middle Atlantic Bight (MAB) to collect field measurements in order to develop algorithms to retrieve CDOM and DOC from NASA's MODIS-Aqua and SeaWiFS satellite sensors. In order to develop empirical algorithms for CDOM and DOC, we correlated the CDOM absorption coefficient (a(sub cdom)) with in situ radiometry (remote sensing reflectance, Rrs, band ratios) and then correlated DOC to Rrs band ratios through the CDOM to DOC relationships. Our validation analyses demonstrate successful retrieval of DOC and CDOM from coastal ocean waters using the MODIS-Aqua and SeaWiFS satellite sensors with mean absolute percent differences from field measurements of cdom)(355)1,6 % for a(sub cdom)(443), and 12% for the CDOM spectral slope. To our knowledge, the algorithms presented here represent the first validated algorithms for satellite retrieval of a(sub cdom) DOC, and CDOM spectral slope in the coastal ocean. The satellite-derived DOC and a(sub cdom) products demonstrate the seasonal net ecosystem production of DOC and photooxidation of CDOM from spring to fall. With accurate satellite retrievals of CDOM and DOC, we will be able to apply satellite observations to investigate interannual and decadal-scale variability in surface CDOM and DOC within continental margins and monitor impacts of climate change and anthropogenic activities on coastal ecosystems.
Directory of Open Access Journals (Sweden)
Ren-Shiou Liu
2016-08-01
Full Text Available In this paper, the problem of minimizing electricity cost and the peak system load in smart grids with distributed renewable energy resources is studied. Unlike prior research works that either assume all of the jobs are interruptible or power-shiftable, this paper focuses on more challenging scenarios in which jobs are non-interruptible and non-power-shiftable. In addition, as more and more newly-built homes have rooftop solar arrays, it is assumed that all users are equipped with a solar-plus-battery system in this paper. Thus, power can be drawn from the battery as needed to reduce the cost of electricity or to lower the overall system load. With a quadratic load-dependent cost function, this paper first shows that the electricity cost minimization problem in such a setting is NP-hard and presents a distributed demand-side management algorithm, called DDSM, to solve this. Experimental results show that the proposed DDSM algorithm is effective, scalable and converges to a Nash equilibrium in finite rounds.
Directory of Open Access Journals (Sweden)
Wing Kam Fung
2010-02-01
Full Text Available The case-control study is an important design for testing association between genetic markers and a disease. The Cochran-Armitage trend test (CATT is one of the most commonly used statistics for the analysis of case-control genetic association studies. The asymptotically optimal CATT can be used when the underlying genetic model (mode of inheritance is known. However, for most complex diseases, the underlying genetic models are unknown. Thus, tests robust to genetic model misspecification are preferable to the model-dependant CATT. Two robust tests, MAX3 and the genetic model selection (GMS, were recently proposed. Their asymptotic null distributions are often obtained by Monte-Carlo simulations, because they either have not been fully studied or involve multiple integrations. In this article, we study how components of each robust statistic are correlated, and find a linear dependence among the components. Using this new finding, we propose simple algorithms to calculate asymptotic null distributions for MAX3 and GMS, which greatly reduce the computing intensity. Furthermore, we have developed the R package Rassoc implementing the proposed algorithms to calculate the empirical and asymptotic p values for MAX3 and GMS as well as other commonly used tests in case-control association studies. For illustration, Rassoc is applied to the analysis of case-control data of 17 most significant SNPs reported in four genome-wide association studies.
International Nuclear Information System (INIS)
Carnet, Bernard; Delhumeau, Michel
1971-06-01
The principles of binary analysis applied to the investigation of sequential circuits were used to design a two way coincidence circuit whose input may be, random or periodic variables of constant or variable duration. The output signal strictly reproduces the characteristics of the input signal triggering the coincidence. A coincidence between input signals does not produce any output signal if one of the signals has already triggered the output signal. The characteristics of the output signal in relation to those of the input signal are: minimum time jitter, excellent duration reproducibility and maximum efficiency. Some rules are given for achieving these results. The symmetry, transitivity and non-transitivity characteristics of the edges on the primitive graph are analyzed and lead to some rules for positioning the states on a secondary graph. It is from this graph that the equations of the circuits can be calculated. The development of the circuit and its dynamic testing are discussed. For this testing, the functioning of the circuit is simulated by feeding into the input randomly generated signals
Sequential charged particle reaction
International Nuclear Information System (INIS)
Hori, Jun-ichi; Ochiai, Kentaro; Sato, Satoshi; Yamauchi, Michinori; Nishitani, Takeo
2004-01-01
The effective cross sections for producing the sequential reaction products in F82H, pure vanadium and LiF with respect to the 14.9-MeV neutron were obtained and compared with the estimation ones. Since the sequential reactions depend on the secondary charged particles behavior, the effective cross sections are corresponding to the target nuclei and the material composition. The effective cross sections were also estimated by using the EAF-libraries and compared with the experimental ones. There were large discrepancies between estimated and experimental values. Additionally, we showed the contribution of the sequential reaction on the induced activity and dose rate in the boundary region with water. From the present study, it has been clarified that the sequential reactions are of great importance to evaluate the dose rates around the surface of cooling pipe and the activated corrosion products. (author)
Bulyha, Alena
2011-01-01
In this work, a Monte-Carlo algorithm in the constant-voltage ensemble for the calculation of 3d charge concentrations at charged surfaces functionalized with biomolecules is presented. The motivation for this work is the theoretical understanding of biofunctionalized surfaces in nanowire field-effect biosensors (BioFETs). This work provides the simulation capability for the boundary layer that is crucial in the detection mechanism of these sensors; slight changes in the charge concentration in the boundary layer upon binding of analyte molecules modulate the conductance of nanowire transducers. The simulation of biofunctionalized surfaces poses special requirements on the Monte-Carlo simulations and these are addressed by the algorithm. The constant-voltage ensemble enables us to include the right boundary conditions; the dna strands can be rotated with respect to the surface; and several molecules can be placed in a single simulation box to achieve good statistics in the case of low ionic concentrations relevant in experiments. Simulation results are presented for the leading example of surfaces functionalized with pna and with single- and double-stranded dna in a sodium-chloride electrolyte. These quantitative results make it possible to quantify the screening of the biomolecule charge due to the counter-ions around the biomolecules and the electrical double layer. The resulting concentration profiles show a three-layer structure and non-trivial interactions between the electric double layer and the counter-ions. The numerical results are also important as a reference for the development of simpler screening models. © 2011 The Royal Society of Chemistry.
A Novel Robust Communication Algorithm for Distributed Secondary Control of Islanded MicroGrids
DEFF Research Database (Denmark)
Shafiee, Qobad; Dragicevic, Tomislav; Vasquez, Juan Carlos
2013-01-01
Distributed secondary control (DSC) is a new approach for MicroGrids (MGs) such that frequency, voltage and power regulation is made in each unit locally to avoid using a central controller. Due to the constrained traffic pattern required by the secondary control, it is viable to implement...
Unified Framework for Deriving Simultaneous Equation Algorithms for Water Distribution Networks
The known formulations for steady state hydraulics within looped water distribution networks are re-derived in terms of linear and non-linear transformations of the original set of partly linear and partly non-linear equations that express conservation of mass and energy. All of ...
A distributed control algorithm for internal flow management in a multi-zone climate unit
Persis, C. De; Jessen, J.J.; Izadi-Zamanabadi, R.; Schiøler, H.
2008-01-01
We examine a distributed control problem for internal flow management in a multi-zone climate unit. The problem consists of guaranteeing prescribed indoor climate conditions in a cascade connection of an arbitrarily large number of communicating zones, in which air masses are exchanged to redirect
Energy Technology Data Exchange (ETDEWEB)
Llacer Martos, S.; Puchal Ane, R.
2011-07-01
To characterize the uniformity of a gamma camera extrinsic used integral uniformity coefficient is calculated with the value of two pixels, the maximum and minimum, single source acquisition of a flat and uniform. This method does not take into account the fact that if a gamma camera having a uniform response, the distribution of these items should be random. In this paper we study how these points are distributed in a succession of large numbers of uniform images.
Distributed Market-Based Algorithms for Multi-Agent Planning with Shared Resources
2013-02-01
1 Introduction 1 2 Distributed Market-Based Multi-Agent Planning 5 2.1 Problem Formulation...over the deterministic planner, on the “test set” of scenarios with changing economies. . . 50 xi xii Chapter 1 Introduction Multi-agent planning is...representation of the objective (4.2.1). For example, for the supply chain mangement problem, we assumed a sequence of Bernoulli coin flips, which seems
Savic, Vladimir; Schiller, Elad M.; Papatriantafilou, Marina
2017-01-01
Vehicle-to-vehicle (V2V) communication is a crucial component of the future autonomous driving systems since it enables improved awareness of the surrounding environment, even without extensive processing of sensory information. However, V2V communication is prone to failures and delays, so a distributed fault-tolerant approach is required for safe and efficient transportation. In this paper, we focus on the intersection crossing (IC) problem with autonomous vehicles that cooperate via V2V co...
Short paper: Distributed vehicular traffic congestion detection algorithm for urban environments
Milojevic, M.; Rakocevic, V.
2013-01-01
Vehicular traffic congestion is a well-known economic and social problem generating significant costs and safety challenges, and increasing pollution in the cities. Current intelligent transport systems and vehicular networking technologies rely heavily on the supporting network infrastructure which is still not widely available. This paper contributes towards the development of distributed and cooperative vehicular traffic congestion detection by proposing a new vehicle-to-vehicle (V2V) cong...
k-means algorithm and mixture distributions for locating faults in power systems
Energy Technology Data Exchange (ETDEWEB)
Mora-Florez, J. [The Technological University of Pereira, La Julita, Ciudad Universitaria, Pereira, Risaralda (Colombia); Cormane-Angarita, J.; Ordonez-Plata, G. [The Industrial University of Santander (Colombia)
2009-05-15
Enhancement of power distribution system reliability requires of a considerable investment in studies and equipment, however, not all the utilities have the capability to spend time and money to assume it. Therefore, any strategy that allows the improvement of reliability should be reflected directly in the reduction of the duration and frequency interruption indexes (SAIFI and SAIDI). In this paper, an alternative solution to the problem of power service continuity associated to fault location is presented. A methodology of statistical nature based on finite mixtures is proposed. A statistical model is obtained from the extraction of the magnitude of the voltage sag registered during a fault event, along with the network parameters and topology. The objective is to offer an economic alternative of easy implementation for the development of strategies oriented to improve the reliability from the reduction of the restoration times in power distribution systems. In the application case for an application example in a power distribution system, the faulted zones were identified, having low error rates. (author)
International Nuclear Information System (INIS)
Sedighizadeh, Mostafa; Esmaili, Masoud; Esmaeili, Mobin
2014-01-01
In this paper, a multi-objective framework is proposed for simultaneous optimal network reconfiguration and DG (distributed generation) power allocation. The proposed method encompasses objective functions of power losses, voltage stability, DG cost, and greenhouse gas emissions and it is optimized subject to power system operational and technical constraints. In order to solve the optimization problem, the HBB-BC (Hybrid Big Bang-Big Crunch) algorithm as one of the most recent heuristic tools is modified and employed here by introducing a mutation operator to enhance its exploration capability. To resolve the scaling problem of differently-scaled objective functions, a fuzzy membership is used to bring them into a same scale and then, the fuzzy fitness of the final objective function is utilized to measure the satisfaction level of the obtained solution. The proposed method is tested on balanced and unbalanced test systems and its results are comprehensively compared with previous methods considering different scenarios. According to results, the proposed method not only offers an enhanced exploration capability but also has a better converge rate compared with previous methods. In addition, the simultaneous network reconfiguration and DG power allocation leads to a more optimal result than separately doing tasks of reconfiguration and DG power allocation. - Highlights: • Hybrid Big Bang-Big Crunch algorithm is applied to network reconfiguration problem. • Joint reconfiguration and DG power allocation leads to a more optimal solution. • A mutation operator is used to improve the exploration capability of HBB-BC method. • The HBB-BC has a better convergence rate than the compared algorithms
Network Reduction Algorithm for Developing Distribution Feeders for Real-Time Simulators: Preprint
Energy Technology Data Exchange (ETDEWEB)
Nagarajan, Adarsh; Nelson, Austin; Prabakar, Kumaraguru; Hoke, Andy; Asano, Marc; Ueda, Reid; Nepal, Shaili
2017-06-15
As advanced grid-support functions (AGF) become more widely used in grid-connected photovoltaic (PV) inverters, utilities are increasingly interested in their impacts when implemented in the field. These effects can be understood by modeling feeders in real-time systems and testing PV inverters using power hardware-in-the-loop (PHIL) techniques. This paper presents a novel feeder model reduction algorithm using a Monte Carlo method that enables large feeders to be solved and operated on real-time computing platforms. Two Hawaiian Electric feeder models in Synergi Electric's load flow software were converted to reduced order models in OpenDSS, and subsequently implemented in the OPAL-RT real-time digital testing platform. Smart PV inverters were added to the real-time model with AGF responses modeled after characterizing commercially available hardware inverters. Finally, hardware inverters were tested in conjunction with the real-time model using PHIL techniques so that the effects of AGFs on the choice feeders could be analyzed.
Xu, Sai; Zhou, Zhiyan; Lu, Huazhong; Luo, Xiwen; Lan, Yubin
2014-03-19
Principal Component Analysis (PCA) is one of the main methods used for electronic nose pattern recognition. However, poor classification performance is common in classification and recognition when using regular PCA. This paper aims to improve the classification performance of regular PCA based on the existing Wilks Λ-statistic (i.e., combined PCA with the Wilks distribution). The improved algorithms, which combine regular PCA with the Wilks Λ-statistic, were developed after analysing the functionality and defects of PCA. Verification tests were conducted using a PEN3 electronic nose. The collected samples consisted of the volatiles of six varieties of rough rice (Zhongxiang1, Xiangwan13, Yaopingxiang, WufengyouT025, Pin 36, and Youyou122), grown in same area and season. The first two principal components used as analysis vectors cannot perform the rough rice varieties classification task based on a regular PCA. Using the improved algorithms, which combine the regular PCA with the Wilks Λ-statistic, many different principal components were selected as analysis vectors. The set of data points of the Mahalanobis distance between each of the varieties of rough rice was selected to estimate the performance of the classification. The result illustrates that the rough rice varieties classification task is achieved well using the improved algorithm. A Probabilistic Neural Networks (PNN) was also established to test the effectiveness of the improved algorithms. The first two principal components (namely PC1 and PC2) and the first and fifth principal component (namely PC1 and PC5) were selected as the inputs of PNN for the classification of the six rough rice varieties. The results indicate that the classification accuracy based on the improved algorithm was improved by 6.67% compared to the results of the regular method. These results prove the effectiveness of using the Wilks Λ-statistic to improve the classification accuracy of the regular PCA approach. The results
Stochastic Evaluation of Maximum Wind Installation in a Radial Distribution Network
DEFF Research Database (Denmark)
Chen, Peiyuan; Bak-Jensen, Birgitte; Chen, Zhe
2011-01-01
This paper proposes an optimization algorithm to find the maximum wind installation in a radial distribution network. The algorithm imposes a limit on the amount of wind energy that can be curtailed annually. The algorithm implements the wind turbine reactive power control and wind energy...... curtailment using sensitivity factors. The optimization is integrated with Monte Carlo simulation to account for the stochastic behavior of load demand and wind power generation. The proposed algorithm is tested on a real 20 kV Danish distribution system in Støvring. It is demonstrated that the algorithm...... executes reactive compensation and energy curtailment sequentially in an effective and efficient manner....
Alsolami, Fawaz; Chikalov, Igor; Moshkov, Mikhail
2013-01-01
This paper is devoted to the study of algorithms for sequential optimization of approximate inhibitory rules relative to the length, coverage and number of misclassifications. Theses algorithms are based on extensions of dynamic programming approach
Sequential Optimization of Paths in Directed Graphs Relative to Different Cost Functions
Abubeker, Jewahir Ali; Chikalov, Igor; Hussain, Shahid; Moshkov, Mikhail
2011-01-01
This paper is devoted to the consideration of an algorithm for sequential optimization of paths in directed graphs relative to di_erent cost functions. The considered algorithm is based on an extension of dynamic programming which allows
Real-time distributed scheduling algorithm for supporting QoS over WDM networks
Kam, Anthony C.; Siu, Kai-Yeung
1998-10-01
Most existing or proposed WDM networks employ circuit switching, typically with one session having exclusive use of one entire wavelength. Consequently they are not suitable for data applications involving bursty traffic patterns. The MIT AON Consortium has developed an all-optical LAN/MAN testbed which provides time-slotted WDM service and employs fast-tunable transceivers in each optical terminal. In this paper, we explore extensions of this service to achieve fine-grained statistical multiplexing with different virtual circuits time-sharing the wavelengths in a fair manner. In particular, we develop a real-time distributed protocol for best-effort traffic over this time-slotted WDM service with near-optical fairness and throughput characteristics. As an additional design feature, our protocol supports the allocation of guaranteed bandwidths to selected connections. This feature acts as a first step towards supporting integrated services and quality-of-service guarantees over WDM networks. To achieve high throughput, our approach is based on scheduling transmissions, as opposed to collision- based schemes. Our distributed protocol involves one MAN scheduler and several LAN schedulers (one per LAN) in a master-slave arrangement. Because of propagation delays and limits on control channel capacities, all schedulers are designed to work with partial, delayed traffic information. Our distributed protocol is of the `greedy' type to ensure fast execution in real-time in response to dynamic traffic changes. It employs a hybrid form of rate and credit control for resource allocation. We have performed extensive simulations, which show that our protocol allocates resources (transmitters, receivers, wavelengths) fairly with high throughput, and supports bandwidth guarantees.
A distributed algorithm for demand-side management: Selling back to the grid.
Latifi, Milad; Khalili, Azam; Rastegarnia, Amir; Zandi, Sajad; Bazzi, Wael M
2017-11-01
Demand side energy consumption scheduling is a well-known issue in the smart grid research area. However, there is lack of a comprehensive method to manage the demand side and consumer behavior in order to obtain an optimum solution. The method needs to address several aspects, including the scale-free requirement and distributed nature of the problem, consideration of renewable resources, allowing consumers to sell electricity back to the main grid, and adaptivity to a local change in the solution point. In addition, the model should allow compensation to consumers and ensurance of certain satisfaction levels. To tackle these issues, this paper proposes a novel autonomous demand side management technique which minimizes consumer utility costs and maximizes consumer comfort levels in a fully distributed manner. The technique uses a new logarithmic cost function and allows consumers to sell excess electricity (e.g. from renewable resources) back to the grid in order to reduce their electric utility bill. To develop the proposed scheme, we first formulate the problem as a constrained convex minimization problem. Then, it is converted to an unconstrained version using the segmentation-based penalty method. At each consumer location, we deploy an adaptive diffusion approach to obtain the solution in a distributed fashion. The use of adaptive diffusion makes it possible for consumers to find the optimum energy consumption schedule with a small number of information exchanges. Moreover, the proposed method is able to track drifts resulting from changes in the price parameters and consumer preferences. Simulations and numerical results show that our framework can reduce the total load demand peaks, lower the consumer utility bill, and improve the consumer comfort level.
A distributed algorithm for demand-side management: Selling back to the grid
Directory of Open Access Journals (Sweden)
Milad Latifi
2017-11-01
Full Text Available Demand side energy consumption scheduling is a well-known issue in the smart grid research area. However, there is lack of a comprehensive method to manage the demand side and consumer behavior in order to obtain an optimum solution. The method needs to address several aspects, including the scale-free requirement and distributed nature of the problem, consideration of renewable resources, allowing consumers to sell electricity back to the main grid, and adaptivity to a local change in the solution point. In addition, the model should allow compensation to consumers and ensurance of certain satisfaction levels. To tackle these issues, this paper proposes a novel autonomous demand side management technique which minimizes consumer utility costs and maximizes consumer comfort levels in a fully distributed manner. The technique uses a new logarithmic cost function and allows consumers to sell excess electricity (e.g. from renewable resources back to the grid in order to reduce their electric utility bill. To develop the proposed scheme, we first formulate the problem as a constrained convex minimization problem. Then, it is converted to an unconstrained version using the segmentation-based penalty method. At each consumer location, we deploy an adaptive diffusion approach to obtain the solution in a distributed fashion. The use of adaptive diffusion makes it possible for consumers to find the optimum energy consumption schedule with a small number of information exchanges. Moreover, the proposed method is able to track drifts resulting from changes in the price parameters and consumer preferences. Simulations and numerical results show that our framework can reduce the total load demand peaks, lower the consumer utility bill, and improve the consumer comfort level. Keywords: Energy, Systems engineering, Electrical engineering
Tanjung, WN; Nurhasanah, N.; Suri, QA; Jingga; Aribowo, B.; Mardhika, DA; Gayatri, AM; Safitri, R.; Supriyanto, A.
2017-12-01
The textile industry is one of the 10 commodities of industrial products which are still survives in Indonesia due to the crisis in the year 2009 until 2016. Drawback happened in 2017 by increased the number of demand by approximate 3% compares with previous year. In this case, the research conducted in Small Medium Enterprise (SME) called FBS. SME is a business group that is able to absorb a lot of labor and a source of income for society. SME FBS producing clothing boys and domiciled in Jakarta. To complete FBS product, the WIP products are sent to CMT or depot in Sukabumi. In this study, aims to do the shortest route in the determination of the distribution of WIP product to 10 CMT scattered in the area of Sukabumi. After optimization hapened, the route must be started from the Depot SME FBS Sukabumi-Shell Sand Village - village of Sukamaju Village - Margaluyu Village - Narogong Cicurug, - the village of Parakanlima, Cuguha - Padabeunghar Village - Sagaranten Village - village of Ciherang, Ciguyang, Sagaranten - the village of Bojong Waru, Pasirsalam Village, Purabaya - Students - return to Depot SME FBS Sukabumi with mileage in a single trip of 403.6 kilometers. It spents 10 hours 09 minutes and cost distribution issued amounting to IDR 296,928.52. The route length was optimized 47% from 759.1 become 403.6 kilometers.
Energy Technology Data Exchange (ETDEWEB)
Calderaro, V.; Coppola, V.; Galdi, V.; Piccolo, A. [Salerno Univ., Fisciano (Italy). Dept. of Information System Engineering and Electrical Engineering
2008-07-01
A new model of power distribution system has emerged in recent years in response to new generation technologies involving mini- and micro-generators that can be directly connected to medium voltage (MV) or low voltage (LV) power grids. The locations of these dispersed generators (DGs) are typically based on the availability of primary energy resources or on the specific needs of users. The increasing use of DGs causes new problems in terms of distribution network management and planning, with effect on the power quality, voltage profile or protection aspects. One of the problems arising on MV/LV distribution network, especially in weak rural areas, is related to the bus overvoltage at the point of common coupling (PCC). Therefore, this study proposed an approach to power control of the single generator that maximizes the active power injected on the network by DG, avoiding the trip of the minimum and maximum voltage protection installed at the PCC. Overvoltage typically occurs due to the injection of a large amount of power from unschedulable DG and a small power demand by the loads. This can trip overvoltage protection relays of DGs, and disconnect them from the grid. The local control strategy for DG systems proposed in this paper was based on the dithering algorithm. The proposed solution, operating on the electronic interface of the power generator, introduces or absorbs reactive power if the voltage at PCC is close to the limits, thus increasing the total active power injected by renewable sources. 17 refs., 3 tabs., 12 figs.
Arpaia, P; Cimmino, P; Girone, M; La Commara, G; Maisto, D; Manna, C; Pezzetti, M
2014-09-01
Evolutionary approach to centralized multiple-faults diagnostics is extended to distributed transducer networks monitoring large experimental systems. Given a set of anomalies detected by the transducers, each instance of the multiple-fault problem is formulated as several parallel communicating sub-tasks running on different transducers, and thus solved one-by-one on spatially separated parallel processes. A micro-genetic algorithm merges evaluation time efficiency, arising from a small-size population distributed on parallel-synchronized processors, with the effectiveness of centralized evolutionary techniques due to optimal mix of exploitation and exploration. In this way, holistic view and effectiveness advantages of evolutionary global diagnostics are combined with reliability and efficiency benefits of distributed parallel architectures. The proposed approach was validated both (i) by simulation at CERN, on a case study of a cold box for enhancing the cryogeny diagnostics of the Large Hadron Collider, and (ii) by experiments, under the framework of the industrial research project MONDIEVOB (Building Remote Monitoring and Evolutionary Diagnostics), co-funded by EU and the company Del Bo srl, Napoli, Italy.
International Nuclear Information System (INIS)
Phuoc, Le Minh; Lee, Suk Han; Kim, Hun Mo; Martinet, Philippe
2008-01-01
Robot inverse kinematics based on Jacobian inversion encounters critical issues of kinematic singularities. In this paper, several techniques based on damped least squares are proposed to lead robot pass through kinematic singularities without excessive joint velocities. Unlike other work in which the same damping factor is used for all singular vectors, this paper proposes a different damping coefficient for each singular vector based on corresponding singular value of the Jacobian. Moreover, a continuous distribution of damping factor following Gaussian function guarantees the continuous in joint velocities. A genetic algorithm is utilized to search for the best maximum damping factor and singular region, which used to require ad hoc searching in other works. As a result, end effector tracking error, which is inherited from damped least squares by introducing damping factors, is minimized. The effectiveness of our approach is compared with other methods in both non-redundant robot and redundant robot
Energy Technology Data Exchange (ETDEWEB)
Phuoc, Le Minh; Lee, Suk Han; Kim, Hun Mo [Sungkyunkwan University, Suwon (Korea, Republic of); Martinet, Philippe [Blaise Pascal University, Clermont-Ferrand Cedex (France)
2008-07-15
Robot inverse kinematics based on Jacobian inversion encounters critical issues of kinematic singularities. In this paper, several techniques based on damped least squares are proposed to lead robot pass through kinematic singularities without excessive joint velocities. Unlike other work in which the same damping factor is used for all singular vectors, this paper proposes a different damping coefficient for each singular vector based on corresponding singular value of the Jacobian. Moreover, a continuous distribution of damping factor following Gaussian function guarantees the continuous in joint velocities. A genetic algorithm is utilized to search for the best maximum damping factor and singular region, which used to require ad hoc searching in other works. As a result, end effector tracking error, which is inherited from damped least squares by introducing damping factors, is minimized. The effectiveness of our approach is compared with other methods in both non-redundant robot and redundant robot
Energy Technology Data Exchange (ETDEWEB)
Ali, I; Algan, O; Ahmad, S [University of Oklahoma Health Sciences, Oklahoma City, OK (United States); Alsbou, N [University of Central Oklahoma, Edmond, OK (United States)
2016-06-15
Purpose: To model patient motion and produce four-dimensional (4D) optimized dose distributions that consider motion-artifacts in the dose calculation during the treatment planning process. Methods: An algorithm for dose calculation is developed where patient motion is considered in dose calculation at the stage of the treatment planning. First, optimal dose distributions are calculated for the stationary target volume where the dose distributions are optimized considering intensity-modulated radiation therapy (IMRT). Second, a convolution-kernel is produced from the best-fitting curve which matches the motion trajectory of the patient. Third, the motion kernel is deconvolved with the initial dose distribution optimized for the stationary target to produce a dose distribution that is optimized in four-dimensions. This algorithm is tested with measured doses using a mobile phantom that moves with controlled motion patterns. Results: A motion-optimized dose distribution is obtained from the initial dose distribution of the stationary target by deconvolution with the motion-kernel of the mobile target. This motion-optimized dose distribution is equivalent to that optimized for the stationary target using IMRT. The motion-optimized and measured dose distributions are tested with the gamma index with a passing rate of >95% considering 3% dose-difference and 3mm distance-to-agreement. If the dose delivery per beam takes place over several respiratory cycles, then the spread-out of the dose distributions is only dependent on the motion amplitude and not affected by motion frequency and phase. This algorithm is limited to motion amplitudes that are smaller than the length of the target along the direction of motion. Conclusion: An algorithm is developed to optimize dose in 4D. Besides IMRT that provides optimal dose coverage for a stationary target, it extends dose optimization to 4D considering target motion. This algorithm provides alternative to motion management
Heterogeneous architecture to process swarm optimization algorithms
Directory of Open Access Journals (Sweden)
Maria A. Dávila-Guzmán
2014-01-01
Full Text Available Since few years ago, the parallel processing has been embedded in personal computers by including co-processing units as the graphics processing units resulting in a heterogeneous platform. This paper presents the implementation of swarm algorithms on this platform to solve several functions from optimization problems, where they highlight their inherent parallel processing and distributed control features. In the swarm algorithms, each individual and dimension problem are parallelized by the granularity of the processing system which also offer low communication latency between individuals through the embedded processing. To evaluate the potential of swarm algorithms on graphics processing units we have implemented two of them: the particle swarm optimization algorithm and the bacterial foraging optimization algorithm. The algorithms’ performance is measured using the acceleration where they are contrasted between a typical sequential processing platform and the NVIDIA GeForce GTX480 heterogeneous platform; the results show that the particle swarm algorithm obtained up to 36.82x and the bacterial foraging swarm algorithm obtained up to 9.26x. Finally, the effect to increase the size of the population is evaluated where we show both the dispersion and the quality of the solutions are decreased despite of high acceleration performance since the initial distribution of the individuals can converge to local optimal solution.
Sequential pattern recognition by maximum conditional informativity
Czech Academy of Sciences Publication Activity Database
Grim, Jiří
2014-01-01
Roč. 45, č. 1 (2014), s. 39-45 ISSN 0167-8655 R&D Projects: GA ČR(CZ) GA14-02652S; GA ČR(CZ) GA14-10911S Keywords : Multivariate statistics * Statistical pattern recognition * Sequential decision making * Product mixtures * EM algorithm * Shannon information Subject RIV: IN - Informatics, Computer Sci ence Impact factor: 1.551, year: 2014 http://library.utia.cas.cz/separaty/2014/RO/grim-0428565.pdf
Directory of Open Access Journals (Sweden)
Shaofeng Wang
2017-05-01
Full Text Available Mineral reserve estimation and mining design depend on a precise modeling of the mineralized deposit. A multi-step interpolation algorithm, including 1D biharmonic spline estimator for interpolating floor altitudes, 2D nearest neighbor, linear, natural neighbor, cubic, biharmonic spline, inverse distance weighted, simple kriging, and ordinary kriging interpolations for grade distribution on the two vertical sections at roadways, and 3D linear interpolation for grade distribution between sections, was proposed to build a 3D grade distribution model of the mineralized seam in a longwall mining panel with a U-shaped layout having two roadways at both sides. Compared to field data from exploratory boreholes, this multi-step interpolation using a natural neighbor method shows an optimal stability and a minimal difference between interpolation and field data. Using this method, the 97,576 m3 of bauxite, in which the mass fraction of Al2O3 (Wa and the mass ratio of Al2O3 to SiO2 (Wa/s are 61.68% and 27.72, respectively, was delimited from the 189,260 m3 mineralized deposit in the 1102 longwall mining panel in the Wachangping mine, Southwest China. The mean absolute errors, the root mean squared errors and the relative standard deviations of errors between interpolated data and exploratory grade data at six boreholes are 2.544, 2.674, and 32.37% of Wa; and 1.761, 1.974, and 67.37% of Wa/s, respectively. The proposed method can be used for characterizing the grade distribution in a mineralized seam between two roadways at both sides of a longwall mining panel.
Sequential stochastic optimization
Cairoli, Renzo
1996-01-01
Sequential Stochastic Optimization provides mathematicians and applied researchers with a well-developed framework in which stochastic optimization problems can be formulated and solved. Offering much material that is either new or has never before appeared in book form, it lucidly presents a unified theory of optimal stopping and optimal sequential control of stochastic processes. This book has been carefully organized so that little prior knowledge of the subject is assumed; its only prerequisites are a standard graduate course in probability theory and some familiarity with discrete-paramet
Inverse problems with non-trivial priors: efficient solution through sequential Gibbs sampling
DEFF Research Database (Denmark)
Hansen, Thomas Mejer; Cordua, Knud Skou; Mosegaard, Klaus
2012-01-01
Markov chain Monte Carlo methods such as the Gibbs sampler and the Metropolis algorithm can be used to sample solutions to non-linear inverse problems. In principle, these methods allow incorporation of prior information of arbitrary complexity. If an analytical closed form description of the prior...... is available, which is the case when the prior can be described by a multidimensional Gaussian distribution, such prior information can easily be considered. In reality, prior information is often more complex than can be described by the Gaussian model, and no closed form expression of the prior can be given....... We propose an algorithm, called sequential Gibbs sampling, allowing the Metropolis algorithm to efficiently incorporate complex priors into the solution of an inverse problem, also for the case where no closed form description of the prior exists. First, we lay out the theoretical background...
Osler, James Edward
2016-01-01
This paper provides a novel instructional methodology that is a unique E-Learning engineered "4A Metric Algorithm" designed to conceptually address the four main challenges faced by 21st century students, who are tempted to cheat in a myriad of higher education settings (face to face, hybrid, and online). The algorithmic online…
Search for the algorithm of genes distribution during the process of microbial evolution
Pikuta, Elena V.
2015-09-01
Previous two and three dimensional graph analysis of eco-physiological data of Archaea demonstrated specific geometry for distribution of major Prokaryotic groups in a hyperboloid function. The function of a two-sheet hyperboloid covered all known biological groups, and therefore, could be applied for the entire evolution of life on Earth. The vector of evolution was indicated from the point of hyper temperature, extreme acidity and low salinity to the point of low temperature and increased alkalinity and salinity. According to this vector, the following groups were chosen for the gene screening analysis. In the vector "High-Temperature → Low-Temperature" within extreme acidic pH (0-3), it is: 1) the hyperthermophilic Crenarchaeota - order Sulfolobales, 2) moderately thermophilic Euryarchaeota - Class Thermoplasmata, and 3) mesophilic acidophiles- genus Thiobacillus and others. In the vector "Low pH → High pH" the following groups were selected in three temperature ranges: a) Hyperthermophilic Archaea and Eubacteria, b) moderately thermophilic - representatives of the genera Anaerobacter and Anoxybacillus, and c) mesophilic haloalkaliphiles (Eubacteria and Archaea). The genes associated with acidophily (H+ pump), chemolitho-autotrophy (proteins of biochemichal cycles), polymerases, and histones were proposed for the first vector, and for the second vector the genes associated with halo-alkaliphily (Na+ pumps), enzymes of organotrophic metabolisms (sugar- and proteolytics), and others were indicated for the screening. Here, an introduction to the phylogenetic constant (ρη) is presented and discussed. This universal characteristic is calculated for two principally different life forms -Prokaryotes and Eukaryotes; Existence of the second type of living forms is impossible without the first one. The number of chromosomes in Prokaryotic organisms is limited to one (with very rare exceptions, to two), while in Eukaryotic organisms this number is larger. Currently
Reinharz, Vladimir; Ponty, Yann; Waldispühl, Jérôme
2013-07-01
The design of RNA sequences folding into predefined secondary structures is a milestone for many synthetic biology and gene therapy studies. Most of the current software uses similar local search strategies (i.e. a random seed is progressively adapted to acquire the desired folding properties) and more importantly do not allow the user to control explicitly the nucleotide distribution such as the GC-content in their sequences. However, the latter is an important criterion for large-scale applications as it could presumably be used to design sequences with better transcription rates and/or structural plasticity. In this article, we introduce IncaRNAtion, a novel algorithm to design RNA sequences folding into target secondary structures with a predefined nucleotide distribution. IncaRNAtion uses a global sampling approach and weighted sampling techniques. We show that our approach is fast (i.e. running time comparable or better than local search methods), seedless (we remove the bias of the seed in local search heuristics) and successfully generates high-quality sequences (i.e. thermodynamically stable) for any GC-content. To complete this study, we develop a hybrid method combining our global sampling approach with local search strategies. Remarkably, our glocal methodology overcomes both local and global approaches for sampling sequences with a specific GC-content and target structure. IncaRNAtion is available at csb.cs.mcgill.ca/incarnation/. Supplementary data are available at Bioinformatics online.
Sequential and Simultaneous Logit: A Nested Model.
van Ophem, J.C.M.; Schram, A.J.H.C.
1997-01-01
A nested model is presented which has both the sequential and the multinomial logit model as special cases. This model provides a simple test to investigate the validity of these specifications. Some theoretical properties of the model are discussed. In the analysis a distribution function is
A Bayesian sequential design with adaptive randomization for 2-sided hypothesis test.
Yu, Qingzhao; Zhu, Lin; Zhu, Han
2017-11-01
Bayesian sequential and adaptive randomization designs are gaining popularity in clinical trials thanks to their potentials to reduce the number of required participants and save resources. We propose a Bayesian sequential design with adaptive randomization rates so as to more efficiently attribute newly recruited patients to different treatment arms. In this paper, we consider 2-arm clinical trials. Patients are allocated to the 2 arms with a randomization rate to achieve minimum variance for the test statistic. Algorithms are presented to calculate the optimal randomization rate, critical values, and power for the proposed design. Sensitivity analysis is implemented to check the influence on design by changing the prior distributions. Simulation studies are applied to compare the proposed method and traditional methods in terms of power and actual sample sizes. Simulations show that, when total sample size is fixed, the proposed design can obtain greater power and/or cost smaller actual sample size than the traditional Bayesian sequential design. Finally, we apply the proposed method to a real data set and compare the results with the Bayesian sequential design without adaptive randomization in terms of sample sizes. The proposed method can further reduce required sample size. Copyright © 2017 John Wiley & Sons, Ltd.
Sequential memory: Binding dynamics
Afraimovich, Valentin; Gong, Xue; Rabinovich, Mikhail
2015-10-01
Temporal order memories are critical for everyday animal and human functioning. Experiments and our own experience show that the binding or association of various features of an event together and the maintaining of multimodality events in sequential order are the key components of any sequential memories—episodic, semantic, working, etc. We study a robustness of binding sequential dynamics based on our previously introduced model in the form of generalized Lotka-Volterra equations. In the phase space of the model, there exists a multi-dimensional binding heteroclinic network consisting of saddle equilibrium points and heteroclinic trajectories joining them. We prove here the robustness of the binding sequential dynamics, i.e., the feasibility phenomenon for coupled heteroclinic networks: for each collection of successive heteroclinic trajectories inside the unified networks, there is an open set of initial points such that the trajectory going through each of them follows the prescribed collection staying in a small neighborhood of it. We show also that the symbolic complexity function of the system restricted to this neighborhood is a polynomial of degree L - 1, where L is the number of modalities.
Sequential Dependencies in Driving
Doshi, Anup; Tran, Cuong; Wilder, Matthew H.; Mozer, Michael C.; Trivedi, Mohan M.
2012-01-01
The effect of recent experience on current behavior has been studied extensively in simple laboratory tasks. We explore the nature of sequential effects in the more naturalistic setting of automobile driving. Driving is a safety-critical task in which delayed response times may have severe consequences. Using a realistic driving simulator, we find…
Mining compressing sequential problems
Hoang, T.L.; Mörchen, F.; Fradkin, D.; Calders, T.G.K.
2012-01-01
Compression based pattern mining has been successfully applied to many data mining tasks. We propose an approach based on the minimum description length principle to extract sequential patterns that compress a database of sequences well. We show that mining compressing patterns is NP-Hard and
Visual Appearance-Based Unmanned Vehicle Sequential Localization
Directory of Open Access Journals (Sweden)
Wei Liu
2013-01-01
Full Text Available Localizationis of vital importance for an unmanned vehicle to drive on the road. Most of the existing algorithms are based on laser range finders, inertial equipment, artificial landmarks, distributing sensors or global positioning system(GPS information. Currently, the problem of localization with vision information is most concerned. However, vision-based localization techniquesare still unavailable for practical applications. In this paper, we present a vision-based sequential probability localization method. This method uses the surface information of the roadside to locate the vehicle, especially in the situation where GPS information is unavailable. It is composed of two step, first, in a recording stage, we construct a ground truthmap with the appearance of the roadside environment. Then in an on-line stage, we use a sequential matching approach to localize the vehicle. In the experiment, we use two independent cameras to observe the environment, one is left-orientated and the other is right. SIFT features and Daisy features are used to represent for the visual appearance of the environment. The experiment results show that the proposed method could locate the vehicle in a complicated, large environment with high reliability.
Li, Cheng
Wind farms, photovoltaic arrays, fuel cells, and micro-turbines are all considered to be Distributed Generation (DG). DG is defined as the generation of power which is dispersed throughout a utility's service territory and either connected to the utility's distribution system or isolated in a small grid. This thesis addresses modeling and economic issues pertaining to the optimal reactive power planning for distribution system with wind power generation (WPG) units. Wind farms are inclined to cause reverse power flows and voltage variations due to the random-like outputs of wind turbines. To deal with this kind of problem caused by wide spread usage of wind power generation, this thesis investigates voltage and reactive power controls in such a distribution system. Consequently static capacitors (SC) and transformer taps are introduced into the system and treated as controllers. For the purpose of getting optimum voltage and realizing reactive power control, the research proposes a proper coordination among the controllers like on-load tap changer (OLTC), feeder-switched capacitors. What's more, in order to simulate its uncertainty, the wind power generation is modeled by the Markov model. In that way, calculating the probabilities for all the scenarios is possible. Some outputs with consecutive and discrete values have been used for transition between successive time states and within state wind speeds. The thesis will describe the method to generate the wind speed time series from the transition probability matrix. After that, utilizing genetic algorithm, the optimal locations of SCs, the sizes of SCs and transformer taps are determined so as to minimize the cost or minimize the power loss, and more importantly improve voltage profiles. The applicability of the proposed method is verified through simulation on a 9-bus system and a 30-bus system respectively. At last, the simulation results indicate that as long as the available capacitors are able to sufficiently
International Nuclear Information System (INIS)
Xu, Meng; Droguett, Enrique López; Lins, Isis Didier; Chagas Moura, Márcio das
2017-01-01
The q-Weibull model is based on the Tsallis non-extensive entropy and is able to model various behaviors of the hazard rate function, including bathtub curves, by using a single set of parameters. Despite its flexibility, the q-Weibull has not been widely used in reliability applications partly because of the complicated parameters estimation. In this work, the parameters of the q-Weibull are estimated by the maximum likelihood (ML) method. Due to the intricate system of nonlinear equations, derivative-based optimization methods may fail to converge. Thus, the heuristic optimization method of artificial bee colony (ABC) is used instead. To deal with the slow convergence of ABC, it is proposed an adaptive hybrid ABC (AHABC) algorithm that dynamically combines Nelder-Mead simplex search method with ABC for the ML estimation of the q-Weibull parameters. Interval estimates for the q-Weibull parameters, including confidence intervals based on the ML asymptotic theory and on bootstrap methods, are also developed. The AHABC is validated via numerical experiments involving the q-Weibull ML for reliability applications and results show that it produces faster and more accurate convergence when compared to ABC and similar approaches. The estimation procedure is applied to real reliability failure data characterized by a bathtub-shaped hazard rate. - Highlights: • Development of an Adaptive Hybrid ABC (AHABC) algorithm for q-Weibull distribution. • AHABC combines local Nelder-Mead simplex method with ABC to enhance local search. • AHABC efficiently finds the optimal solution for the q-Weibull ML problem. • AHABC outperforms ABC and self-adaptive hybrid ABC in accuracy and convergence speed. • Useful model for reliability data with non-monotonic hazard rate.
International Nuclear Information System (INIS)
Kefayat, M.; Lashkar Ara, A.; Nabavi Niaki, S.A.
2015-01-01
Highlights: • A probabilistic optimization framework incorporated with uncertainty is proposed. • A hybrid optimization approach combining ACO and ABC algorithms is proposed. • The problem is to deal with technical, environmental and economical aspects. • A fuzzy interactive approach is incorporated to solve the multi-objective problem. • Several strategies are implemented to compare with literature methods. - Abstract: In this paper, a hybrid configuration of ant colony optimization (ACO) with artificial bee colony (ABC) algorithm called hybrid ACO–ABC algorithm is presented for optimal location and sizing of distributed energy resources (DERs) (i.e., gas turbine, fuel cell, and wind energy) on distribution systems. The proposed algorithm is a combined strategy based on the discrete (location optimization) and continuous (size optimization) structures to achieve advantages of the global and local search ability of ABC and ACO algorithms, respectively. Also, in the proposed algorithm, a multi-objective ABC is used to produce a set of non-dominated solutions which store in the external archive. The objectives consist of minimizing power losses, total emissions produced by substation and resources, total electrical energy cost, and improving the voltage stability. In order to investigate the impact of the uncertainty in the output of the wind energy and load demands, a probabilistic load flow is necessary. In this study, an efficient point estimate method (PEM) is employed to solve the optimization problem in a stochastic environment. The proposed algorithm is tested on the IEEE 33- and 69-bus distribution systems. The results demonstrate the potential and effectiveness of the proposed algorithm in comparison with those of other evolutionary optimization methods
Algorithm for Compressing Time-Series Data
Hawkins, S. Edward, III; Darlington, Edward Hugo
2012-01-01
An algorithm based on Chebyshev polynomials effects lossy compression of time-series data or other one-dimensional data streams (e.g., spectral data) that are arranged in blocks for sequential transmission. The algorithm was developed for use in transmitting data from spacecraft scientific instruments to Earth stations. In spite of its lossy nature, the algorithm preserves the information needed for scientific analysis. The algorithm is computationally simple, yet compresses data streams by factors much greater than two. The algorithm is not restricted to spacecraft or scientific uses: it is applicable to time-series data in general. The algorithm can also be applied to general multidimensional data that have been converted to time-series data, a typical example being image data acquired by raster scanning. However, unlike most prior image-data-compression algorithms, this algorithm neither depends on nor exploits the two-dimensional spatial correlations that are generally present in images. In order to understand the essence of this compression algorithm, it is necessary to understand that the net effect of this algorithm and the associated decompression algorithm is to approximate the original stream of data as a sequence of finite series of Chebyshev polynomials. For the purpose of this algorithm, a block of data or interval of time for which a Chebyshev polynomial series is fitted to the original data is denoted a fitting interval. Chebyshev approximation has two properties that make it particularly effective for compressing serial data streams with minimal loss of scientific information: The errors associated with a Chebyshev approximation are nearly uniformly distributed over the fitting interval (this is known in the art as the "equal error property"); and the maximum deviations of the fitted Chebyshev polynomial from the original data have the smallest possible values (this is known in the art as the "min-max property").
Directory of Open Access Journals (Sweden)
Byongjun Hwang
2017-07-01
Full Text Available In this study, we present an algorithm for summer sea ice conditions that semi-automatically produces the floe size distribution of Arctic sea ice from high-resolution satellite Synthetic Aperture Radar data. Currently, floe size distribution data from satellite images are very rare in the literature, mainly due to the lack of a reliable algorithm to produce such data. Here, we developed the algorithm by combining various image analysis methods, including Kernel Graph Cuts, distance transformation and watershed transformation, and a rule-based boundary revalidation. The developed algorithm has been validated against the ground truth that was extracted manually with the aid of 1-m resolution visible satellite data. Comprehensive validation analysis has shown both perspectives and limitations. The algorithm tends to fail to detect small floes (mostly less than 100 m in mean caliper diameter compared to ground truth, which is mainly due to limitations in water-ice segmentation. Some variability in the power law exponent of floe size distribution is observed due to the effects of control parameters in the process of de-noising, Kernel Graph Cuts segmentation, thresholds for boundary revalidation and image resolution. Nonetheless, the algorithm, for floes larger than 100 m, has shown a reasonable agreement with ground truth under various selections of these control parameters. Considering that the coverage and spatial resolution of satellite Synthetic Aperture Radar data have increased significantly in recent years, the developed algorithm opens a new possibility to produce large volumes of floe size distribution data, which is essential for improving our understanding and prediction of the Arctic sea ice cover
A Survey of Multi-Objective Sequential Decision-Making
Roijers, D.M.; Vamplew, P.; Whiteson, S.; Dazeley, R.
2013-01-01
Sequential decision-making problems with multiple objectives arise naturally in practice and pose unique challenges for research in decision-theoretic planning and learning, which has largely focused on single-objective settings. This article surveys algorithms designed for sequential decision-making problems with multiple objectives. Though there is a growing body of literature on this subject, little of it makes explicit under what circumstances special methods are needed to solve multi-obj...
Directory of Open Access Journals (Sweden)
Asma Shaheen
2018-03-01
Full Text Available In third world countries, industries mainly cause environmental contamination due to lack of environmental policies or oversight during their implementation. The Sheikhupura industrial zone, which includes industries such as tanneries, leather, chemical, textiles, and colour and dyes, contributes massive amounts of untreated effluents that are released directly into drains and used for the irrigation of crops and vegetables. This practice causes not only soil contamination with an excessive amount of heavy metals, but is also considered a source of toxicity in the food chain, i.e., bioaccumulation in plants and ultimately in human body organs. The objective of this research study was to assess the spatial distribution of the heavy metals chromium (Cr, cadmium (Cd, and lead (Pb, at three depths of soil using geostatistics and the selection of significant contributing variables to soil contamination using the Random Forest (RF function of the Boruta Algorithm. A total of 60 sampling locations were selected in the study area to collect soil samples (180 samples at three depths (0–15 cm, 15–30 cm, and 60–90 cm. The soil samples were analysed for their physico-chemical properties, i.e., soil saturation, electrical conductivity (EC, organic matter (OM, pH, phosphorus (P, potassium (K, and Cr, Cd, and Pb using standard laboratory procedures. The data were analysed with comprehensive statistics and geostatistical techniques. The correlation coefficient matrix between the heavy metals and the physico-chemical properties revealed that electrical conductivity (EC had a significant (p ≤ 0.05 negative correlation with Cr, Cd, and Pb. The RF function of the Boruta Algorithm employed soil depth as a classifier and ranked the significant soil contamination parameters (Cr, Cd, Pb, EC, and P in relation to depth. The mobility factor indicated the leachate percentage of heavy metals at different vertical depths of soil. The spatial distribution pattern of
Ma, B.; Li, J.; Fan, W.; Ren, H.; Xu, X.
2017-12-01
Leaf area index (LAI) is one of the important parameters of vegetation canopy structure, which can represent the growth condition of vegetation effectively. The accuracy, availability and timeliness of LAI data can be improved greatly, which is of great importance to vegetation-related research, such as the study of atmospheric, land surface and hydrological processes to obtain LAI by remote sensing method. Heihe River Basin is the inland river basin in northwest China. There are various types of vegetation and all kinds of terrain conditions in the basin, so it is helpful for testing the accuracy of the model under the complex surface and evaluating the correctness of the model to study LAI in this area. On the other hand, located in west arid area of China, the ecological environment of Heihe Basin is fragile, LAI is an important parameter to represent the vegetation growth condition, and can help us understand the status of vegetation in the Heihe River Basin. Different from the previous LAI inversion models, the BRDF (bidirectional reflectance distribution function) unified model can be applied for both continuous vegetation and discrete vegetation, it is appropriate to the complex vegetation distribution. LAI is the key input parameter of the model. We establish the inversion algorithm that can exactly retrieve LAI using remote sensing image based on the unified model. First, we determine the vegetation type through the vegetation classification map to obtain the corresponding G function, leaf and surface reflectivity. Then, we need to determine the leaf area index (LAI), the aggregation index (ζ) and the sky scattered light ratio (β) range and the value of the interval, entering all the parameters into the model to calculate the corresponding reflectivity ρ and establish the lookup table of different vegetation. Finally, we can invert LAI on the basis of the established lookup table. The principle of inversion is least squares method. We have produced 1 km
Energy Technology Data Exchange (ETDEWEB)
Niknam, Taher; Meymand, Hamed Zeinoddini; Mojarrad, Hasan Doagou [Department of Electrical and Electronics Engineering, Shiraz University of Technology, Shiraz, P.O. 71555-313 (Iran, Islamic Republic of)
2011-05-15
In this paper a novel Multi-objective fuzzy self adaptive hybrid particle swarm optimization (MFSAHPSO) evolutionary algorithm to solve the Multi-objective optimal operation management (MOOM) is presented. The purposes of the MOOM problem are to decrease the total electrical energy losses, the total electrical energy cost and the total pollutant emission produced by fuel cells and substation bus. Conventional algorithms used to solve the multi-objective optimization problems convert the multiple objectives into a single objective, using a vector of the user-predefined weights. In this conversion several deficiencies can be detected. For instance, the optimal solution of the algorithms depends greatly on the values of the weights and also some of the information may be lost in the conversion process and so this strategy is not expected to provide a robust solution. This paper presents a new MFSAHPSO algorithm for the MOOM problem. The proposed algorithm maintains a finite-sized repository of non-dominated solutions which gets iteratively updated in the presence of new solutions. Since the objective functions are not the same, a fuzzy clustering technique is used to control the size of the repository, within the limits. The proposed algorithm is tested on a distribution test feeder and the results demonstrate the capabilities of the proposed approach, to generate true and well-distributed Pareto-optimal non-dominated solutions of the MOOM problem. (author)
Forced Sequence Sequential Decoding
DEFF Research Database (Denmark)
Jensen, Ole Riis
In this thesis we describe a new concatenated decoding scheme based on iterations between an inner sequentially decoded convolutional code of rate R=1/4 and memory M=23, and block interleaved outer Reed-Solomon codes with non-uniform profile. With this scheme decoding with good performance...... is possible as low as Eb/No=0.6 dB, which is about 1.7 dB below the signal-to-noise ratio that marks the cut-off rate for the convolutional code. This is possible since the iteration process provides the sequential decoders with side information that allows a smaller average load and minimizes the probability...... of computational overflow. Analytical results for the probability that the first Reed-Solomon word is decoded after C computations are presented. This is supported by simulation results that are also extended to other parameters....
Papaconstadopoulos, P; Levesque, I R; Maglieri, R; Seuntjens, J
2016-02-07
Direct determination of the source intensity distribution of clinical linear accelerators is still a challenging problem for small field beam modeling. Current techniques most often involve special equipment and are difficult to implement in the clinic. In this work we present a maximum-likelihood expectation-maximization (MLEM) approach to the source reconstruction problem utilizing small fields and a simple experimental set-up. The MLEM algorithm iteratively ray-traces photons from the source plane to the exit plane and extracts corrections based on photon fluence profile measurements. The photon fluence profiles were determined by dose profile film measurements in air using a high density thin foil as build-up material and an appropriate point spread function (PSF). The effect of other beam parameters and scatter sources was minimized by using the smallest field size ([Formula: see text] cm(2)). The source occlusion effect was reproduced by estimating the position of the collimating jaws during this process. The method was first benchmarked against simulations for a range of typical accelerator source sizes. The sources were reconstructed with an accuracy better than 0.12 mm in the full width at half maximum (FWHM) to the respective electron sources incident on the target. The estimated jaw positions agreed within 0.2 mm with the expected values. The reconstruction technique was also tested against measurements on a Varian Novalis Tx linear accelerator and compared to a previously commissioned Monte Carlo model. The reconstructed FWHM of the source agreed within 0.03 mm and 0.11 mm to the commissioned electron source in the crossplane and inplane orientations respectively. The impact of the jaw positioning, experimental and PSF uncertainties on the reconstructed source distribution was evaluated with the former presenting the dominant effect.
Directory of Open Access Journals (Sweden)
Wen-Xiang Wu
2014-01-01
Full Text Available The cost-based system optimum problem in networks with continuously distributed value of time is formulated as a path-based form, which cannot be solved by the Frank-Wolfe algorithm. In light of magnitude improvement in the availability of computer memory in recent years, path-based algorithms have been regarded as a viable approach for traffic assignment problems with reasonably large network sizes. We develop a path-based gradient projection algorithm for solving the cost-based system optimum model, based on Goldstein-Levitin-Polyak method which has been successfully applied to solve standard user equilibrium and system optimum problems. The Sioux Falls network tested is used to verify the effectiveness of the algorithm.
Sequential Power-Dependence Theory
Buskens, Vincent; Rijt, Arnout van de
2008-01-01
Existing methods for predicting resource divisions in laboratory exchange networks do not take into account the sequential nature of the experimental setting. We extend network exchange theory by considering sequential exchange. We prove that Sequential Power-Dependence Theory—unlike
Modelling sequentially scored item responses
Akkermans, W.
2000-01-01
The sequential model can be used to describe the variable resulting from a sequential scoring process. In this paper two more item response models are investigated with respect to their suitability for sequential scoring: the partial credit model and the graded response model. The investigation is
Directory of Open Access Journals (Sweden)
N. Bayati
2017-02-01
Full Text Available Distributed Generation (DG connection in a power system tends to increase the short circuit level in the entire system which, in turn, could eliminate the protection coordination between the existing relays. Fault Current Limiters (FCLs are often used to reduce the short-circuit level of the network to a desirable level, provided that they are dully placed and appropriately sized. In this paper, a method is proposed for optimal placement of FCLs and optimal determination of their impedance values by which the relay operation time, the number and size of the FCL are minimized while maintaining the relay coordination before and after DG connection. The proposed method adopts the removal of low-impact FCLs and uses a hybrid Genetic Algorithm (GA optimization scheme to determine the optimal placement of FCLs and the values of their impedances. The suitability of the proposed method is demonstrated by examining the results of relay coordination in a typical DG network before and after DG connection.
Forced Sequence Sequential Decoding
DEFF Research Database (Denmark)
Jensen, Ole Riis; Paaske, Erik
1998-01-01
We describe a new concatenated decoding scheme based on iterations between an inner sequentially decoded convolutional code of rate R=1/4 and memory M=23, and block interleaved outer Reed-Solomon (RS) codes with nonuniform profile. With this scheme decoding with good performance is possible as low...... as Eb/N0=0.6 dB, which is about 1.25 dB below the signal-to-noise ratio (SNR) that marks the cutoff rate for the full system. Accounting for about 0.45 dB due to the outer codes, sequential decoding takes place at about 1.7 dB below the SNR cutoff rate for the convolutional code. This is possible since...... the iteration process provides the sequential decoders with side information that allows a smaller average load and minimizes the probability of computational overflow. Analytical results for the probability that the first RS word is decoded after C computations are presented. These results are supported...
Random and cooperative sequential adsorption
Evans, J. W.
1993-10-01
Irreversible random sequential adsorption (RSA) on lattices, and continuum "car parking" analogues, have long received attention as models for reactions on polymer chains, chemisorption on single-crystal surfaces, adsorption in colloidal systems, and solid state transformations. Cooperative generalizations of these models (CSA) are sometimes more appropriate, and can exhibit richer kinetics and spatial structure, e.g., autocatalysis and clustering. The distribution of filled or transformed sites in RSA and CSA is not described by an equilibrium Gibbs measure. This is the case even for the saturation "jammed" state of models where the lattice or space cannot fill completely. However exact analysis is often possible in one dimension, and a variety of powerful analytic methods have been developed for higher dimensional models. Here we review the detailed understanding of asymptotic kinetics, spatial correlations, percolative structure, etc., which is emerging for these far-from-equilibrium processes.
Optimization algorithms and applications
Arora, Rajesh Kumar
2015-01-01
Choose the Correct Solution Method for Your Optimization ProblemOptimization: Algorithms and Applications presents a variety of solution techniques for optimization problems, emphasizing concepts rather than rigorous mathematical details and proofs. The book covers both gradient and stochastic methods as solution techniques for unconstrained and constrained optimization problems. It discusses the conjugate gradient method, Broyden-Fletcher-Goldfarb-Shanno algorithm, Powell method, penalty function, augmented Lagrange multiplier method, sequential quadratic programming, method of feasible direc
Marker-controlled watershed for lymphoma segmentation in sequential CT images
International Nuclear Information System (INIS)
Yan Jiayong; Zhao Binsheng; Wang, Liang; Zelenetz, Andrew; Schwartz, Lawrence H.
2006-01-01
Segmentation of lymphoma containing lymph nodes is a difficult task because of multiple variables associated with the tumor's location, intensity distribution, and contrast to its surrounding tissues. In this paper, we present a reliable and practical marker-controlled watershed algorithm for semi-automated segmentation of lymphoma in sequential CT images. Robust determination of internal and external markers is the key to successful use of the marker-controlled watershed transform in the segmentation of lymphoma and is the focus of this work. The external marker in our algorithm is the circle enclosing the lymphoma in a single slice. The internal marker, however, is determined automatically by combining techniques including Canny edge detection, thresholding, morphological operation, and distance map estimation. To obtain tumor volume, the segmented lymphoma in the current slice needs to be propagated to the adjacent slice to help determine the external and internal markers for delineation of the lymphoma in that slice. The algorithm was applied to 29 lymphomas (size range, 9-53 mm in diameter; mean, 23 mm) in nine patients. A blinded radiologist manually delineated all lymphomas on all slices. The manual result served as the ''gold standard'' for comparison. Several quantitative methods were applied to objectively evaluate the performance of the segmentation algorithm. The algorithm received a mean overlap, overestimation, and underestimation ratios of 83.2%, 13.5%, and 5.5%, respectively. The mean average boundary distance and Hausdorff boundary distance were 0.7 and 3.7 mm. Preliminary results have shown the potential of this computer algorithm to allow reliable segmentation and quantification of lymphomas on sequential CT images
Automatic synthesis of sequential control schemes
International Nuclear Information System (INIS)
Klein, I.
1993-01-01
Of all hard- and software developed for industrial control purposes, the majority is devoted to sequential, or binary valued, control and only a minor part to classical linear control. Typically, the sequential parts of the controller are invoked during startup and shut-down to bring the system into its normal operating region and into some safe standby region, respectively. Despite its importance, fairly little theoretical research has been devoted to this area, and sequential control programs are therefore still created manually without much theoretical support to obtain a systematic approach. We propose a method to create sequential control programs automatically. The main ideas is to spend some effort off-line modelling the plant, and from this model generate the control strategy, that is the plan. The plant is modelled using action structures, thereby concentrating on the actions instead of the states of the plant. In general the planning problem shows exponential complexity in the number of state variables. However, by focusing on the actions, we can identify problem classes as well as algorithms such that the planning complexity is reduced to polynomial complexity. We prove that these algorithms are sound, i.e., the generated solution will solve the stated problem, and complete, i.e., if the algorithms fail, then no solution exists. The algorithms generate a plan as a set of actions and a partial order on this set specifying the execution order. The generated plant is proven to be minimal and maximally parallel. For a larger class of problems we propose a method to split the original problem into a number of simple problems that can each be solved using one of the presented algorithms. It is also shown how a plan can be translated into a GRAFCET chart, and to illustrate these ideas we have implemented a planing tool, i.e., a system that is able to automatically create control schemes. Such a tool can of course also be used on-line if it is fast enough. This
Multilevel sequential Monte Carlo samplers
Beskos, Alexandros; Jasra, Ajay; Law, Kody; Tempone, Raul; Zhou, Yan
2016-01-01
In this article we consider the approximation of expectations w.r.t. probability distributions associated to the solution of partial differential equations (PDEs); this scenario appears routinely in Bayesian inverse problems. In practice, one often has to solve the associated PDE numerically, using, for instance finite element methods which depend on the step-size level . hL. In addition, the expectation cannot be computed analytically and one often resorts to Monte Carlo methods. In the context of this problem, it is known that the introduction of the multilevel Monte Carlo (MLMC) method can reduce the amount of computational effort to estimate expectations, for a given level of error. This is achieved via a telescoping identity associated to a Monte Carlo approximation of a sequence of probability distributions with discretization levels . âˆž>h0>h1â‹¯>hL. In many practical problems of interest, one cannot achieve an i.i.d. sampling of the associated sequence and a sequential Monte Carlo (SMC) version of the MLMC method is introduced to deal with this problem. It is shown that under appropriate assumptions, the attractive property of a reduction of the amount of computational effort to estimate expectations, for a given level of error, can be maintained within the SMC context. That is, relative to exact sampling and Monte Carlo for the distribution at the finest level . hL. The approach is numerically illustrated on a Bayesian inverse problem. Â© 2016 Elsevier B.V.
Multilevel sequential Monte Carlo samplers
Beskos, Alexandros
2016-08-29
In this article we consider the approximation of expectations w.r.t. probability distributions associated to the solution of partial differential equations (PDEs); this scenario appears routinely in Bayesian inverse problems. In practice, one often has to solve the associated PDE numerically, using, for instance finite element methods which depend on the step-size level . hL. In addition, the expectation cannot be computed analytically and one often resorts to Monte Carlo methods. In the context of this problem, it is known that the introduction of the multilevel Monte Carlo (MLMC) method can reduce the amount of computational effort to estimate expectations, for a given level of error. This is achieved via a telescoping identity associated to a Monte Carlo approximation of a sequence of probability distributions with discretization levels . âˆž>h0>h1â‹¯>hL. In many practical problems of interest, one cannot achieve an i.i.d. sampling of the associated sequence and a sequential Monte Carlo (SMC) version of the MLMC method is introduced to deal with this problem. It is shown that under appropriate assumptions, the attractive property of a reduction of the amount of computational effort to estimate expectations, for a given level of error, can be maintained within the SMC context. That is, relative to exact sampling and Monte Carlo for the distribution at the finest level . hL. The approach is numerically illustrated on a Bayesian inverse problem. Â© 2016 Elsevier B.V.
International Nuclear Information System (INIS)
Yoshida, Toshihiro
1981-01-01
Probabilities of meson production in the sequential decay of Reggeons, which are formed from the projectile and the target in the hadron-hadron to Reggeon-Reggeon processes, are investigated. It is assumed that pair creation of heavy quarks and simultaneous creation of two antiquark-quark pairs are negligible. The leading-order terms with respect to ratio of creation probabilities of anti s s to anti u u (anti d d) are calculated. The production cross sections in the target fragmentation region are given in terms of probabilities in the initial decay of the Reggeons and an effect of manyparticle production. (author)
International Nuclear Information System (INIS)
Choi, H; Shin, W; Testa, M; Min, C; Kim, J
2015-01-01
Purpose: For intensity-modulated radiation therapy (IMRT) treatment planning validation using Monte Carlo (MC) simulations, a precise and automated procedure is necessary to evaluate the patient dose distribution. The aim of this study is to develop an automated algorithm for IMRT simulations using DICOM files and to evaluate the patient dose based on 4D simulation using the Geant4 MC toolkit. Methods: The head of a clinical linac (Varian Clinac 2300 IX) was modeled in Geant4 along with particular components such as the flattening filter and the multi-leaf collimator (MLC). Patient information and the position of the MLC were imported from the DICOM-RT interface. For each position of the MLC, a step- and-shoot technique was adopted. PDDs and lateral profiles were simulated in a water phantom (50×50×40 cm 3 ) and compared to measurement data. We used a lung phantom and MC-dose calculations were compared to the clinical treatment planning used at the Seoul National University Hospital. Results: In order to reproduce the measurement data, we tuned three free parameters: mean and standard deviation of the primary electron beam energy and the beam spot size. These parameters for 6 MV were found to be 5.6 MeV, 0.2378 MeV and 1 mm FWHM respectively. The average dose difference between measurements and simulations was less than 2% for PDDs and radial profiles. The lung phantom study showed fairly good agreement between MC and planning dose despite some unavoidable statistical fluctuation. Conclusion: The current feasibility study using the lung phantom shows the potential for IMRT dose validation using 4D MC simulations using Geant4 tool kits. This research was supported by Korea Institute of Nuclear safety and Development of Measurement Standards for Medical Radiation funded by Korea research Institute of Standards and Science. (KRISS-2015-15011032)
Yavari, Somayeh; Valadan Zoej, Mohammad Javad; Salehi, Bahram
2018-05-01
The procedure of selecting an optimum number and best distribution of ground control information is important in order to reach accurate and robust registration results. This paper proposes a new general procedure based on Genetic Algorithm (GA) which is applicable for all kinds of features (point, line, and areal features). However, linear features due to their unique characteristics are of interest in this investigation. This method is called Optimum number of Well-Distributed ground control Information Selection (OWDIS) procedure. Using this method, a population of binary chromosomes is randomly initialized. The ones indicate the presence of a pair of conjugate lines as a GCL and zeros specify the absence. The chromosome length is considered equal to the number of all conjugate lines. For each chromosome, the unknown parameters of a proper mathematical model can be calculated using the selected GCLs (ones in each chromosome). Then, a limited number of Check Points (CPs) are used to evaluate the Root Mean Square Error (RMSE) of each chromosome as its fitness value. The procedure continues until reaching a stopping criterion. The number and position of ones in the best chromosome indicate the selected GCLs among all conjugate lines. To evaluate the proposed method, a GeoEye and an Ikonos Images are used over different areas of Iran. Comparing the obtained results by the proposed method in a traditional RFM with conventional methods that use all conjugate lines as GCLs shows five times the accuracy improvement (pixel level accuracy) as well as the strength of the proposed method. To prevent an over-parametrization error in a traditional RFM due to the selection of a high number of improper correlated terms, an optimized line-based RFM is also proposed. The results show the superiority of the combination of the proposed OWDIS method with an optimized line-based RFM in terms of increasing the accuracy to better than 0.7 pixel, reliability, and reducing systematic
The Bacterial Sequential Markov Coalescent.
De Maio, Nicola; Wilson, Daniel J
2017-05-01
Bacteria can exchange and acquire new genetic material from other organisms directly and via the environment. This process, known as bacterial recombination, has a strong impact on the evolution of bacteria, for example, leading to the spread of antibiotic resistance across clades and species, and to the avoidance of clonal interference. Recombination hinders phylogenetic and transmission inference because it creates patterns of substitutions (homoplasies) inconsistent with the hypothesis of a single evolutionary tree. Bacterial recombination is typically modeled as statistically akin to gene conversion in eukaryotes, i.e. , using the coalescent with gene conversion (CGC). However, this model can be very computationally demanding as it needs to account for the correlations of evolutionary histories of even distant loci. So, with the increasing popularity of whole genome sequencing, the need has emerged for a faster approach to model and simulate bacterial genome evolution. We present a new model that approximates the coalescent with gene conversion: the bacterial sequential Markov coalescent (BSMC). Our approach is based on a similar idea to the sequential Markov coalescent (SMC)-an approximation of the coalescent with crossover recombination. However, bacterial recombination poses hurdles to a sequential Markov approximation, as it leads to strong correlations and linkage disequilibrium across very distant sites in the genome. Our BSMC overcomes these difficulties, and shows a considerable reduction in computational demand compared to the exact CGC, and very similar patterns in simulated data. We implemented our BSMC model within new simulation software FastSimBac. In addition to the decreased computational demand compared to previous bacterial genome evolution simulators, FastSimBac provides more general options for evolutionary scenarios, allowing population structure with migration, speciation, population size changes, and recombination hotspots. FastSimBac is
Online co-regularized algorithms
Ruijter, T. de; Tsivtsivadze, E.; Heskes, T.
2012-01-01
We propose an online co-regularized learning algorithm for classification and regression tasks. We demonstrate that by sequentially co-regularizing prediction functions on unlabeled data points, our algorithm provides improved performance in comparison to supervised methods on several UCI benchmarks
Synthetic Aperture Sequential Beamforming
DEFF Research Database (Denmark)
Kortbek, Jacob; Jensen, Jørgen Arendt; Gammelmark, Kim Løkke
2008-01-01
A synthetic aperture focusing (SAF) technique denoted Synthetic Aperture Sequential Beamforming (SASB) suitable for 2D and 3D imaging is presented. The technique differ from prior art of SAF in the sense that SAF is performed on pre-beamformed data contrary to channel data. The objective is to im......A synthetic aperture focusing (SAF) technique denoted Synthetic Aperture Sequential Beamforming (SASB) suitable for 2D and 3D imaging is presented. The technique differ from prior art of SAF in the sense that SAF is performed on pre-beamformed data contrary to channel data. The objective...... is to improve and obtain a more range independent lateral resolution compared to conventional dynamic receive focusing (DRF) without compromising frame rate. SASB is a two-stage procedure using two separate beamformers. First a set of Bmode image lines using a single focal point in both transmit and receive...... is stored. The second stage applies the focused image lines from the first stage as input data. The SASB method has been investigated using simulations in Field II and by off-line processing of data acquired with a commercial scanner. The performance of SASB with a static image object is compared with DRF...
Sequential Computerized Mastery Tests--Three Simulation Studies
Wiberg, Marie
2006-01-01
A simulation study of a sequential computerized mastery test is carried out with items modeled with the 3 parameter logistic item response theory model. The examinees' responses are either identically distributed, not identically distributed, or not identically distributed together with estimation errors in the item characteristics. The…