An Efficient Character Segmentation Based on VNP Algorithm
S. Chitrakala
2012-12-01
Full Text Available Character segmentation is an important preprocessing stage in image processing applications such as OCR, License Plate Recognition, electronic processing of checks in banks, form processing and, label and barcode recognition. It is essential to have an efficient character segmentation technique because it affects the performance of all the processes that follow and hence, the overall system accuracy. Vertical projection profile is the most common segmentation technique. However, the segmentation results are not always correct in cases where pixels of adjacent characters fall on the same scan line and a minimum threshold is not observed in the histogram to segment the respective adjacent characters. In this study, a character segmentation technique based on Visited Neighbor Pixel (VNP Algorithm is proposed, which is an improvement to the vertical projection profile technique. VNP Algorithm performs segmentation based on the connectedness of the pixels on the scan line with that of the previously visited pixels. Therefore, a clear line of separation is found even when the threshold between two adjacent characters is not minimal. The segmentation results of the traditional vertical projection profile and the proposed method are compared with respect to a few selected fonts and the latter, with an average accuracy of approximately 94%, has shown encouraging results.
T. Karpagam
2012-01-01
Full Text Available Problem statement: Network topology design problems find application in several real life scenario. Approach: Most designs in the past either optimize for a single criterion like shortest or cost minimization or maximum flow. Results: This study discussed about solving a multi objective network topology design problem for a realistic traffic model specifically in the pipeline transportation. Here flow based algorithm focusing to transport liquid goods with maximum capacity with shortest distance, this algorithm developed with the sense of basic pert and critical path method. Conclusion/Recommendations: This flow based algorithm helps to give optimal result for transporting maximum capacity with minimum cost. It could be used in the juice factory, milk industry and its best alternate for the vehicle routing problem.
HISTORY BASED PROBABILISTIC BACKOFF ALGORITHM
Narendran Rajagopalan
2012-01-01
Full Text Available Performance of Wireless LAN can be improved at each layer of the protocol stack with respect to energy efficiency. The Media Access Control layer is responsible for the key functions like access control and flow control. During contention, Backoff algorithm is used to gain access to the medium with minimum probability of collision. After studying different variations of back off algorithms that have been proposed, a new variant called History based Probabilistic Backoff Algorithm is proposed. Through mathematical analysis and simulation results using NS-2, it is seen that proposed History based Probabilistic Backoff algorithm performs better than Binary Exponential Backoff algorithm.
Variables Bounding Based Retiming Algorithm
宫宗伟; 林争辉; 陈后鹏
2002-01-01
Retiming is a technique for optimizing sequential circuits. In this paper, wediscuss this problem and propose an improved retiming algorithm based on variables bounding.Through the computation of the lower and upper bounds on variables, the algorithm can signi-ficantly reduce the number of constraints and speed up the execution of retiming. Furthermore,the elements of matrixes D and W are computed in a demand-driven way, which can reducethe capacity of memory. It is shown through the experimental results on ISCAS89 benchmarksthat our algorithm is very effective for large-scale sequential circuits.
Evolutionary algorithm based index assignment algorithm for noisy channel
李天昊; 余松煜
2004-01-01
A globally optimal solution to vector quantization (VQ) index assignment on noisy channel, the evolutionary algorithm based index assignment algorithm (EAIAA), is presented. The algorithm yields a significant reduction in average distortion due to channel errors, over conventional arbitrary index assignment, as confirmed by experimental results over the memoryless binary symmetric channel (BSC) for any bit error.
Miller, Anthony B; Harirchi, Iraj; Lotfi, Mohammad Hassan; Noori, Mahmoud; Mirzaei, Mohsen; Jafarizadea, Majid; Sadeghian, Mohammad Reza; Minosepehr, Mojgan
2015-11-01
There is some evidence to suggest that a benefit might be derived from a program that incorporated both annual physical examination of the breast (BPx) and the teaching of breast self-examination (BSE). Current investigation presents the profile of a multicenter community based intervention for evaluating the effect of BSE+BPx on the reduction of morbidity and mortality due to breast cancer amongst women residing in urban areas of Yazd (Iran) from 2008 to 2018. There were three distinctive phases in this trial with 10 years duration: pilot phase with the duration of 1 year, active intervention phase with 4 rounds of annual screening of BPx+BSE and follow up phase with 5 years duration. Tools of enquiry included a pre-tested questionnaire, repeated annual physical examination of the breast and more importantly mammography, sonography, and fine needle aspiration (FNA). Data were analyzed using descriptive statistics such as frequencies, percent, mean (SD), tests of chi-square and student t-test with 95% confidence level. Comparison of socio-demographic and socio-economic factors such as age, age at marriage, family size, number of live births, occupation, education level, total family income and marital status showed that no significant difference was seen between the groups (P>0.05). A response rate of 84.5% was seen by participants of the experiment group visiting the health centers for the first BPx. Our results showed that except for the education and marital status, the difference in other main demographic and socio-economic factors between the groups were not significant, and the response rate of individuals in the experiment group was at an acceptable level.
Anthony B Miller
2015-11-01
Full Text Available There is some evidence to suggest that a benefit might be derived from a program that incorporated both annual physical examination of the breast (BPx and the teaching of breast self-examination (BSE. Current investigation presents the profile of a multicenter community based intervention for evaluating the effect of BSE+BPx on the reduction of morbidity and mortality due to breast cancer amongst women residing in urban areas of Yazd (Iran from 2008 to 2018. There were three distinctive phases in this trial with 10 years duration: pilot phase with the duration of 1 year, active intervention phase with 4 rounds of annual screening of BPx+BSE and follow up phase with 5 years duration. Tools of enquiry included a pre-tested questionnaire, repeated annual physical examination of the breast and more importantly mammography, sonography, and fine needle aspiration (FNA. Data were analyzed using descriptive statistics such as frequencies, percent, mean (SD, tests of chi-square and student t-test with 95% confidence level. Comparison of socio-demographic and socio-economic factors such as age, age at marriage, family size, number of live births, occupation, education level, total family income and marital status showed that no significant difference was seen between the groups (P>0.05. A response rate of 84.5% was seen by participants of the experiment group visiting the health centers for the first BPx. Our results showed that except for the education and marital status, the difference in other main demographic and socio-economic factors between the groups were not significant, and the response rate of individuals in the experiment group was at an acceptable level.
Application of detecting algorithm based on network
张凤斌; 杨永田; 江子扬; 孙冰心
2004-01-01
Because currently intrusion detection systems cannot detect undefined intrusion behavior effectively,according to the robustness and adaptability of the genetic algorithms, this paper integrates the genetic algorithms into an intrusion detection system, and a detection algorithm based on network traffic is proposed. This algorithm is a real-time and self-study algorithm and can detect undefined intrusion behaviors effectively.
Diversity-Based Boosting Algorithm
Jafar A. Alzubi
2016-05-01
Full Text Available Boosting is a well known and efficient technique for constructing a classifier ensemble. An ensemble is built incrementally by altering the distribution of training data set and forcing learners to focus on misclassification errors. In this paper, an improvement to Boosting algorithm called DivBoosting algorithm is proposed and studied. Experiments on several data sets are conducted on both Boosting and DivBoosting. The experimental results show that DivBoosting is a promising method for ensemble pruning. We believe that it has many advantages over traditional boosting method because its mechanism is not solely based on selecting the most accurate base classifiers but also based on selecting the most diverse set of classifiers.
AN SVAD ALGORITHM BASED ON FNNKD METHOD
Chen Dong; Zhang Yan; Kuang Jingming
2002-01-01
The capacity of mobile communication system is improved by using Voice Activity Detection (VAD) technology. In this letter, a novel VAD algorithm, SVAD algorithm based on Fuzzy Neural Network Knowledge Discovery (FNNKD) method is proposed. The performance of SVAD algorithm is discussed and compared with traditional algorithm recommended by ITU G.729B in different situations. The simulation results show that the SVAD algorithm performs better.
A New Page Ranking Algorithm Based On WPRVOL Algorithm
Roja Javadian Kootenae
2013-03-01
Full Text Available The amount of information on the web is always growing, thus powerful search tools are needed to search for such a large collection. Search engines in this direction help users so they can find their desirable information among the massive volume of information in an easier way. But what is important in the search engines and causes a distinction between them is page ranking algorithm used in them. In this paper a new page ranking algorithm based on "Weighted Page Ranking based on Visits of Links (WPRVOL Algorithm" for search engines is being proposed which is called WPR'VOL for short. The proposed algorithm considers the number of visits of first and second level in-links. The original WPRVOL algorithm takes into account the number of visits of first level in-links of the pages and distributes rank scores based on the popularity of the pages whereas the proposed algorithm considers both in-links of that page (first level in-links and in-links of the pages that point to it (second level in-links in order to calculation of rank of the page, hence more related pages are displayed at the top of search result list. In the summary it is said that the proposed algorithm assigns higher rank to pages that both themselves and pages that point to them be important.
DNA Coding Based Knowledge Discovery Algorithm
LI Ji-yun; GENG Zhao-feng; SHAO Shi-huang
2002-01-01
A novel DNA coding based knowledge discovery algorithm was proposed, an example which verified its validity was given. It is proved that this algorithm can discover new simplified rules from the original rule set efficiently.
Distance Concentration-Based Artificial Immune Algorithm
LIU Tao; WANG Yao-cai; WANG Zhi-jie; MENG Jiang
2005-01-01
The diversity, adaptation and memory of biological immune system attract much attention of researchers. Several optimal algorithms based on immune system have also been proposed up to now. The distance concentration-based artificial immune algorithm (DCAIA) is proposed to overcome defects of the classical artificial immune algorithm (CAIA) in this paper. Compared with genetic algorithm (GA) and CAIA, DCAIA is good for solving the problem of precocity,holding the diversity of antibody, and enhancing convergence rate.
SIFT based algorithm for point feature tracking
Adrian BURLACU
2007-12-01
Full Text Available In this paper a tracking algorithm for SIFT features in image sequences is developed. For each point feature extracted using SIFT algorithm a descriptor is computed using information from its neighborhood. Using an algorithm based on minimizing the distance between two descriptors tracking point features throughout image sequences is engaged. Experimental results, obtained from image sequences that capture scaling of different geometrical type object, reveal the performances of the tracking algorithm.
Neural Network-Based Hyperspectral Algorithms
2016-06-07
Neural Network-Based Hyperspectral Algorithms Walter F. Smith, Jr. and Juanita Sandidge Naval Research Laboratory Code 7340, Bldg 1105 Stennis Space...our effort is development of robust numerical inversion algorithms , which will retrieve inherent optical properties of the water column as well as...validate the resulting inversion algorithms with in-situ data and provide estimates of the error bounds associated with the inversion algorithm . APPROACH
A New Page Ranking Algorithm Based On WPRVOL Algorithm
Roja Javadian Kootenae; Seyyed Mohsen Hashemi; mehdi afzali
2013-01-01
The amount of information on the web is always growing, thus powerful search tools are needed to search for such a large collection. Search engines in this direction help users so they can find their desirable information among the massive volume of information in an easier way. But what is important in the search engines and causes a distinction between them is page ranking algorithm used in them. In this paper a new page ranking algorithm based on "Weighted Page Ranking based on Visits of ...
Kernel method-based fuzzy clustering algorithm
Wu Zhongdong; Gao Xinbo; Xie Weixin; Yu Jianping
2005-01-01
The fuzzy C-means clustering algorithm(FCM) to the fuzzy kernel C-means clustering algorithm(FKCM) to effectively perform cluster analysis on the diversiform structures are extended, such as non-hyperspherical data, data with noise, data with mixture of heterogeneous cluster prototypes, asymmetric data, etc. Based on the Mercer kernel, FKCM clustering algorithm is derived from FCM algorithm united with kernel method. The results of experiments with the synthetic and real data show that the FKCM clustering algorithm is universality and can effectively unsupervised analyze datasets with variform structures in contrast to FCM algorithm. It is can be imagined that kernel-based clustering algorithm is one of important research direction of fuzzy clustering analysis.
ILU preconditioning based on the FAPINV algorithm
Davod Khojasteh Salkuyeh
2015-01-01
Full Text Available A technique for computing an ILU preconditioner based on the factored approximate inverse (FAPINV algorithm is presented. We show that this algorithm is well-defined for H-matrices. Moreover, when used in conjunction with Krylov-subspace-based iterative solvers such as the GMRES algorithm, results in reliable solvers. Numerical experiments on some test matrices are given to show the efficiency of the new ILU preconditioner.
Multicast Routing Based on Hybrid Genetic Algorithm
CAO Yuan-da; CAI Gui
2005-01-01
A new multicast routing algorithm based on the hybrid genetic algorithm (HGA) is proposed. The coding pattern based on the number of routing paths is used. A fitness function that is computed easily and makes algorithm quickly convergent is proposed. A new approach that defines the HGA's parameters is provided. The simulation shows that the approach can increase largely the convergent ratio, and the fitting values of the parameters of this algorithm are different from that of the original algorithms. The optimal mutation probability of HGA equals 0.50 in HGA in the experiment, but that equals 0.07 in SGA. It has been concluded that the population size has a significant influence on the HGA's convergent ratio when it's mutation probability is bigger. The algorithm with a small population size has a high average convergent rate. The population size has little influence on HGA with the lower mutation probability.
Eigenvalue Decomposition-Based Modified Newton Algorithm
Wen-jun Wang
2013-01-01
Full Text Available When the Hessian matrix is not positive, the Newton direction may not be the descending direction. A new method named eigenvalue decomposition-based modified Newton algorithm is presented, which first takes the eigenvalue decomposition of the Hessian matrix, then replaces the negative eigenvalues with their absolute values, and finally reconstructs the Hessian matrix and modifies the searching direction. The new searching direction is always the descending direction. The convergence of the algorithm is proven and the conclusion on convergence rate is presented qualitatively. Finally, a numerical experiment is given for comparing the convergence domains of the modified algorithm and the classical algorithm.
Immune Based Intrusion Detector Generating Algorithm
DONG Xiao-mei; YU Ge; XIANG Guang
2005-01-01
Immune-based intrusion detection approaches are studied. The methods of constructing self set and generating mature detectors are researched and improved. A binary encoding based self set construction method is applied. First,the traditional mature detector generating algorithm is improved to generate mature detectors and detect intrusions faster. Then, a novel mature detector generating algorithm is proposed based on the negative selection mechanism. Accord ing to the algorithm, less mature detectors are needed to detect the abnormal activities in the network. Therefore, the speed of generating mature detectors and intrusion detection is improved. By comparing with those based on existing algo rithms, the intrusion detection system based on the algorithm has higher speed and accuracy.
Lane Detection Based on Machine Learning Algorithm
Chao Fan; Jingbo Xu; Shuai Di
2013-01-01
In order to improve accuracy and robustness of the lane detection in complex conditions, such as the shadows and illumination changing, a novel detection algorithm was proposed based on machine learning...
QPSO-based adaptive DNA computing algorithm.
Karakose, Mehmet; Cigdem, Ugur
2013-01-01
DNA (deoxyribonucleic acid) computing that is a new computation model based on DNA molecules for information storage has been increasingly used for optimization and data analysis in recent years. However, DNA computing algorithm has some limitations in terms of convergence speed, adaptability, and effectiveness. In this paper, a new approach for improvement of DNA computing is proposed. This new approach aims to perform DNA computing algorithm with adaptive parameters towards the desired goal using quantum-behaved particle swarm optimization (QPSO). Some contributions provided by the proposed QPSO based on adaptive DNA computing algorithm are as follows: (1) parameters of population size, crossover rate, maximum number of operations, enzyme and virus mutation rate, and fitness function of DNA computing algorithm are simultaneously tuned for adaptive process, (2) adaptive algorithm is performed using QPSO algorithm for goal-driven progress, faster operation, and flexibility in data, and (3) numerical realization of DNA computing algorithm with proposed approach is implemented in system identification. Two experiments with different systems were carried out to evaluate the performance of the proposed approach with comparative results. Experimental results obtained with Matlab and FPGA demonstrate ability to provide effective optimization, considerable convergence speed, and high accuracy according to DNA computing algorithm.
QPSO-Based Adaptive DNA Computing Algorithm
Mehmet Karakose
2013-01-01
Full Text Available DNA (deoxyribonucleic acid computing that is a new computation model based on DNA molecules for information storage has been increasingly used for optimization and data analysis in recent years. However, DNA computing algorithm has some limitations in terms of convergence speed, adaptability, and effectiveness. In this paper, a new approach for improvement of DNA computing is proposed. This new approach aims to perform DNA computing algorithm with adaptive parameters towards the desired goal using quantum-behaved particle swarm optimization (QPSO. Some contributions provided by the proposed QPSO based on adaptive DNA computing algorithm are as follows: (1 parameters of population size, crossover rate, maximum number of operations, enzyme and virus mutation rate, and fitness function of DNA computing algorithm are simultaneously tuned for adaptive process, (2 adaptive algorithm is performed using QPSO algorithm for goal-driven progress, faster operation, and flexibility in data, and (3 numerical realization of DNA computing algorithm with proposed approach is implemented in system identification. Two experiments with different systems were carried out to evaluate the performance of the proposed approach with comparative results. Experimental results obtained with Matlab and FPGA demonstrate ability to provide effective optimization, considerable convergence speed, and high accuracy according to DNA computing algorithm.
Evolutionary algorithm based configuration interaction approach
Chakraborty, Rahul
2016-01-01
A stochastic configuration interaction method based on evolutionary algorithm is designed as an affordable approximation to full configuration interaction (FCI). The algorithm comprises of initiation, propagation and termination steps, where the propagation step is performed with cloning, mutation and cross-over, taking inspiration from genetic algorithm. We have tested its accuracy in 1D Hubbard problem and a molecular system (symmetric bond breaking of water molecule). We have tested two different fitness functions based on energy of the determinants and the CI coefficients of determinants. We find that the absolute value of CI coefficients is a more suitable fitness function when combined with a fixed selection scheme.
无
2003-01-01
A new FFT algorithm has been deduced, which is called the base-6 FFT algorithm. The amount for calculating the DFT of complex sequence of N=2r by the base-6 FFT algorithm is Mr(N)=14/3*Nlog6N-4N+4 for multiplication operation of real number and Ar(N)=23/3*Nlog6N-2N+2 for addition operation of real number. The amount for calculating the DFT of real sequence is a half of it with the complex sequence.
Seizure detection algorithms based on EMG signals
Conradsen, Isa
Background: the currently used non-invasive seizure detection methods are not reliable. Muscle fibers are directly connected to the nerves, whereby electric signals are generated during activity. Therefore, an alarm system on electromyography (EMG) signals is a theoretical possibility. Objective......: to show whether medical signal processing of EMG data is feasible for detection of epileptic seizures. Methods: EMG signals during generalised seizures were recorded from 3 patients (with 20 seizures in total). Two possible medical signal processing algorithms were tested. The first algorithm was based...... on the amplitude of the signal. The other algorithm was based on information of the signal in the frequency domain, and it focused on synchronisation of the electrical activity in a single muscle during the seizure. Results: The amplitude-based algorithm reliably detected seizures in 2 of the patients, while...
Duality based optical flow algorithms with applications
Rakêt, Lars Lau
We consider the popular TV-L1 optical flow formulation, and the so-called duality based algorithm for minimizing the TV-L1 energy. The original formulation is extended to allow for vector valued images, and minimization results are given. In addition we consider different definitions of total...... variation regularization, and related formulations of the optical flow problem that may be used with a duality based algorithm. We present a highly optimized algorithmic setup to estimate optical flows, and give five novel applications. The first application is registration of medical images, where X......-ray images of different hands, taken using different imaging devices are registered using a TV-L1 optical flow algorithm. We propose to regularize the input images, using sparsity enhancing regularization of the image gradient to improve registration results. The second application is registration of 2D...
Duality based optical flow algorithms with applications
Rakêt, Lars Lau
We consider the popular TV-L1 optical flow formulation, and the so-called duality based algorithm for minimizing the TV-L1 energy. The original formulation is extended to allow for vector valued images, and minimization results are given. In addition we consider different definitions of total...... variation regularization, and related formulations of the optical flow problem that may be used with a duality based algorithm. We present a highly optimized algorithmic setup to estimate optical flows, and give five novel applications. The first application is registration of medical images, where X......-ray images of different hands, taken using different imaging devices are registered using a TV-L1 optical flow algorithm. We propose to regularize the input images, using sparsity enhancing regularization of the image gradient to improve registration results. The second application is registration of 2D...
Function Optimization Based on Quantum Genetic Algorithm
Ying Sun
2014-01-01
Full Text Available Optimization method is important in engineering design and application. Quantum genetic algorithm has the characteristics of good population diversity, rapid convergence and good global search capability and so on. It combines quantum algorithm with genetic algorithm. A novel quantum genetic algorithm is proposed, which is called Variable-boundary-coded Quantum Genetic Algorithm (vbQGA in which qubit chromosomes are collapsed into variable-boundary-coded chromosomes instead of binary-coded chromosomes. Therefore much shorter chromosome strings can be gained. The method of encoding and decoding of chromosome is first described before a new adaptive selection scheme for angle parameters used for rotation gate is put forward based on the core ideas and principles of quantum computation. Eight typical functions are selected to optimize to evaluate the effectiveness and performance of vbQGA against standard Genetic Algorithm (sGA and Genetic Quantum Algorithm (GQA. The simulation results show that vbQGA is significantly superior to sGA in all aspects and outperforms GQA in robustness and solving velocity, especially for multidimensional and complicated functions.
Edge Crossing Minimization Algorithm for Hierarchical Graphs Based on Genetic Algorithms
无
2001-01-01
We present an edge crossing minimization algorithm forhierarchical gr aphs based on genetic algorithms, and comparing it with some heuristic algorithm s. The proposed algorithm is more efficient and has the following advantages: th e frame of the algorithms is unified, the method is simple, and its implementati on and revision are easy.
Secure OFDM communications based on hashing algorithms
Neri, Alessandro; Campisi, Patrizio; Blasi, Daniele
2007-10-01
In this paper we propose an OFDM (Orthogonal Frequency Division Multiplexing) wireless communication system that introduces mutual authentication and encryption at the physical layer, without impairing spectral efficiency, exploiting some freedom degrees of the base-band signal, and using encrypted-hash algorithms. FEC (Forward Error Correction) is instead performed through variable-rate Turbo Codes. To avoid false rejections, i.e. rejections of enrolled (authorized) users, we designed and tested a robust hash algorithm. This robustness is obtained both by a segmentation of the hash domain (based on BCH codes) and by the FEC capabilities of Turbo Codes.
Graphical model construction based on evolutionary algorithms
Youlong YANG; Yan WU; Sanyang LIU
2006-01-01
Using Bayesian networks to model promising solutions from the current population of the evolutionary algorithms can ensure efficiency and intelligence search for the optimum. However, to construct a Bayesian network that fits a given dataset is a NP-hard problem, and it also needs consuming mass computational resources. This paper develops a methodology for constructing a graphical model based on Bayesian Dirichlet metric. Our approach is derived from a set of propositions and theorems by researching the local metric relationship of networks matching dataset. This paper presents the algorithm to construct a tree model from a set of potential solutions using above approach. This method is important not only for evolutionary algorithms based on graphical models, but also for machine learning and data mining.The experimental results show that the exact theoretical results and the approximations match very well.
A Practical Propositional Knowledge Base Revision Algorithm
陶雪红; 孙伟; 等
1997-01-01
This paper gives an outline of knowledge base revision and some recently presented complexity results about propostitional knowledge base revision.Different methods for revising propositional knowledge base have been proposed recently by several researchers,but all methods are intractable in the general case.For practical application,this paper presents a revision method for special case,and gives its corresponding polynomial algorithm.
Second Attribute Algorithm Based on Tree Expression
Su-Qing Han; Jue Wang
2006-01-01
One view of finding a personalized solution of reduct in an information system is grounded on the viewpoint that attribute order can serve as a kind of semantic representation of user requirements. Thus the problem of finding personalized solutions can be transformed into computing the reduct on an attribute order. The second attribute theorem describes the relationship between the set of attribute orders and the set of reducts, and can be used to transform the problem of searching solutions to meet user requirements into the problem of modifying reduct based on a given attribute order. An algorithm is implied based on the second attribute theorem, with computation on the discernibility matrix. Its time complexity is O(n2 × m) (n is the number of the objects and m the number of the attributes of an information system).This paper presents another effective second attribute algorithm for facilitating the use of the second attribute theorem,with computation on the tree expression of an information system. The time complexity of the new algorithm is linear in n. This algorithm is proved to be equivalent to the algorithm on the discernibility matrix.
Structure-Based Algorithms for Microvessel Classification
Smith, Amy F.
2015-02-01
© 2014 The Authors. Microcirculation published by John Wiley & Sons Ltd. Objective: Recent developments in high-resolution imaging techniques have enabled digital reconstruction of three-dimensional sections of microvascular networks down to the capillary scale. To better interpret these large data sets, our goal is to distinguish branching trees of arterioles and venules from capillaries. Methods: Two novel algorithms are presented for classifying vessels in microvascular anatomical data sets without requiring flow information. The algorithms are compared with a classification based on observed flow directions (considered the gold standard), and with an existing resistance-based method that relies only on structural data. Results: The first algorithm, developed for networks with one arteriolar and one venular tree, performs well in identifying arterioles and venules and is robust to parameter changes, but incorrectly labels a significant number of capillaries as arterioles or venules. The second algorithm, developed for networks with multiple inlets and outlets, correctly identifies more arterioles and venules, but is more sensitive to parameter changes. Conclusions: The algorithms presented here can be used to classify microvessels in large microvascular data sets lacking flow information. This provides a basis for analyzing the distinct geometrical properties and modelling the functional behavior of arterioles, capillaries, and venules.
Model based development of engine control algorithms
Dekker, H.J.; Sturm, W.L.
1996-01-01
Model based development of engine control systems has several advantages. The development time and costs are strongly reduced because much of the development and optimization work is carried out by simulating both engine and control system. After optimizing the control algorithm it can be executed b
Verification-Based Interval-Passing Algorithm for Compressed Sensing
Wu, Xiaofu; Yang, Zhen
2013-01-01
We propose a verification-based Interval-Passing (IP) algorithm for iteratively reconstruction of nonnegative sparse signals using parity check matrices of low-density parity check (LDPC) codes as measurement matrices. The proposed algorithm can be considered as an improved IP algorithm by further incorporation of the mechanism of verification algorithm. It is proved that the proposed algorithm performs always better than either the IP algorithm or the verification algorithm. Simulation resul...
Optimal Hops-Based Adaptive Clustering Algorithm
Xuan, Xin; Chen, Jian; Zhen, Shanshan; Kuo, Yonghong
This paper proposes an optimal hops-based adaptive clustering algorithm (OHACA). The algorithm sets an energy selection threshold before the cluster forms so that the nodes with less energy are more likely to go to sleep immediately. In setup phase, OHACA introduces an adaptive mechanism to adjust cluster head and load balance. And the optimal distance theory is applied to discover the practical optimal routing path to minimize the total energy for transmission. Simulation results show that OHACA prolongs the life of network, improves utilizing rate and transmits more data because of energy balance.
Numerical Algorithms Based on Biorthogonal Wavelets
Ponenti, Pj.; Liandrat, J.
1996-01-01
Wavelet bases are used to generate spaces of approximation for the resolution of bidimensional elliptic and parabolic problems. Under some specific hypotheses relating the properties of the wavelets to the order of the involved operators, it is shown that an approximate solution can be built. This approximation is then stable and converges towards the exact solution. It is designed such that fast algorithms involving biorthogonal multi resolution analyses can be used to resolve the corresponding numerical problems. Detailed algorithms are provided as well as the results of numerical tests on partial differential equations defined on the bidimensional torus.
Algorithmic Differentiation for Calculus-based Optimization
Walther, Andrea
2010-10-01
For numerous applications, the computation and provision of exact derivative information plays an important role for optimizing the considered system but quite often also for its simulation. This presentation introduces the technique of Algorithmic Differentiation (AD), a method to compute derivatives of arbitrary order within working precision. Quite often an additional structure exploitation is indispensable for a successful coupling of these derivatives with state-of-the-art optimization algorithms. The talk will discuss two important situations where the problem-inherent structure allows a calculus-based optimization. Examples from aerodynamics and nano optics illustrate these advanced optimization approaches.
Lane Detection Based on Machine Learning Algorithm
Chao Fan
2013-09-01
Full Text Available In order to improve accuracy and robustness of the lane detection in complex conditions, such as the shadows and illumination changing, a novel detection algorithm was proposed based on machine learning. After pretreatment, a set of haar-like filters were used to calculate the eigenvalue in the gray image f(x,y and edge e(x,y. Then these features were trained by using improved boosting algorithm and the final class function g(x was obtained, which was used to judge whether the point x belonging to the lane or not. To avoid the over fitting in traditional boosting, Fisher discriminant analysis was used to initialize the weights of samples. After testing by many road in all conditions, it showed that this algorithm had good robustness and real-time to recognize the lane in all challenging conditions.
Web Based Genetic Algorithm Using Data Mining
Ashiqur Rahman; Asaduzzaman Noman; Md. Ashraful Islam; Al-Amin Gaji
2016-01-01
This paper presents an approach for classifying students in order to predict their final grade based on features extracted from logged data in an education web-based system. A combination of multiple classifiers leads to a significant improvement in classification performance. Through weighting the feature vectors using a Genetic Algorithm we can optimize the prediction accuracy and get a marked improvement over raw classification. It further shows that when the number of features is few; fea...
AN OPTIMIZATION ALGORITHM BASED ON BACTERIA BEHAVIOR
Ricardo Contreras
2014-09-01
Full Text Available Paradigms based on competition have shown to be useful for solving difficult problems. In this paper we present a new approach for solving hard problems using a collaborative philosophy. A collaborative philosophy can produce paradigms as interesting as the ones found in algorithms based on a competitive philosophy. Furthermore, we show that the performance - in problems associated to explosive combinatorial - is comparable to the performance obtained using a classic evolutive approach.
Cloud-based Evolutionary Algorithms: An algorithmic study
Merelo, Juan-J; Mora, Antonio M; Castillo, Pedro; Romero, Gustavo; Laredo, JLJ
2011-01-01
After a proof of concept using Dropbox(tm), a free storage and synchronization service, showed that an evolutionary algorithm using several dissimilar computers connected via WiFi or Ethernet had a good scaling behavior in terms of evaluations per second, it remains to be proved whether that effect also translates to the algorithmic performance of the algorithm. In this paper we will check several different, and difficult, problems, and see what effects the automatic load-balancing and asynchrony have on the speed of resolution of problems.
A novel tree structure based watermarking algorithm
Lin, Qiwei; Feng, Gui
2008-03-01
In this paper, we propose a new blind watermarking algorithm for images which is based on tree structure. The algorithm embeds the watermark in wavelet transform domain, and the embedding positions are determined by significant coefficients wavelets tree(SCWT) structure, which has the same idea with the embedded zero-tree wavelet (EZW) compression technique. According to EZW concepts, we obtain coefficients that are related to each other by a tree structure. This relationship among the wavelet coefficients allows our technique to embed more watermark data. If the watermarked image is attacked such that the set of significant coefficients is changed, the tree structure allows the correlation-based watermark detector to recover synchronously. The algorithm also uses a visual adaptive scheme to insert the watermark to minimize watermark perceptibility. In addition to the watermark, a template is inserted into the watermarked image at the same time. The template contains synchronization information, allowing the detector to determine the geometric transformations type applied to the watermarked image. Experimental results show that the proposed watermarking algorithm is robust against most signal processing attacks, such as JPEG compression, median filtering, sharpening and rotating. And it is also an adaptive method which shows a good performance to find the best areas to insert a stronger watermark.
Fast Algorithms for Model-Based Diagnosis
Fijany, Amir; Barrett, Anthony; Vatan, Farrokh; Mackey, Ryan
2005-01-01
Two improved new methods for automated diagnosis of complex engineering systems involve the use of novel algorithms that are more efficient than prior algorithms used for the same purpose. Both the recently developed algorithms and the prior algorithms in question are instances of model-based diagnosis, which is based on exploring the logical inconsistency between an observation and a description of a system to be diagnosed. As engineering systems grow more complex and increasingly autonomous in their functions, the need for automated diagnosis increases concomitantly. In model-based diagnosis, the function of each component and the interconnections among all the components of the system to be diagnosed (for example, see figure) are represented as a logical system, called the system description (SD). Hence, the expected behavior of the system is the set of logical consequences of the SD. Faulty components lead to inconsistency between the observed behaviors of the system and the SD. The task of finding the faulty components (diagnosis) reduces to finding the components, the abnormalities of which could explain all the inconsistencies. Of course, the meaningful solution should be a minimal set of faulty components (called a minimal diagnosis), because the trivial solution, in which all components are assumed to be faulty, always explains all inconsistencies. Although the prior algorithms in question implement powerful methods of diagnosis, they are not practical because they essentially require exhaustive searches among all possible combinations of faulty components and therefore entail the amounts of computation that grow exponentially with the number of components of the system.
Research of the Kernel Operator Library Based on Cryptographic Algorithm
王以刚; 钱力; 黄素梅
2001-01-01
The variety of encryption mechanism and algorithms which were conventionally used have some limitations.The kernel operator library based on Cryptographic algorithm is put forward. Owing to the impenetrability of algorithm, the data transfer system with the cryptographic algorithm library has many remarkable advantages in algorithm rebuilding and optimization,easily adding and deleting algorithm, and improving the security power over the traditional algorithm. The user can choose any one in all algorithms with the method against any attack because the cryptographic algorithm library is extensible.
Network-based recommendation algorithms: A review
Yu, Fei; Gillard, Sebastien; Medo, Matus
2015-01-01
Recommender systems are a vital tool that helps us to overcome the information overload problem. They are being used by most e-commerce web sites and attract the interest of a broad scientific community. A recommender system uses data on users' past preferences to choose new items that might be appreciated by a given individual user. While many approaches to recommendation exist, the approach based on a network representation of the input data has gained considerable attention in the past. We review here a broad range of network-based recommendation algorithms and for the first time compare their performance on three distinct real datasets. We present recommendation topics that go beyond the mere question of which algorithm to use - such as the possible influence of recommendation on the evolution of systems that use it - and finally discuss open research directions and challenges.
Network-based recommendation algorithms: A review
Yu, Fei; Zeng, An; Gillard, Sébastien; Medo, Matúš
2016-06-01
Recommender systems are a vital tool that helps us to overcome the information overload problem. They are being used by most e-commerce web sites and attract the interest of a broad scientific community. A recommender system uses data on users' past preferences to choose new items that might be appreciated by a given individual user. While many approaches to recommendation exist, the approach based on a network representation of the input data has gained considerable attention in the past. We review here a broad range of network-based recommendation algorithms and for the first time compare their performance on three distinct real datasets. We present recommendation topics that go beyond the mere question of which algorithm to use-such as the possible influence of recommendation on the evolution of systems that use it-and finally discuss open research directions and challenges.
LSB Based Quantum Image Steganography Algorithm
Jiang, Nan; Zhao, Na; Wang, Luo
2016-01-01
Quantum steganography is the technique which hides a secret message into quantum covers such as quantum images. In this paper, two blind LSB steganography algorithms in the form of quantum circuits are proposed based on the novel enhanced quantum representation (NEQR) for quantum images. One algorithm is plain LSB which uses the message bits to substitute for the pixels' LSB directly. The other is block LSB which embeds a message bit into a number of pixels that belong to one image block. The extracting circuits can regain the secret message only according to the stego cover. Analysis and simulation-based experimental results demonstrate that the invisibility is good, and the balance between the capacity and the robustness can be adjusted according to the needs of applications.
A Single Pattern Matching Algorithm Based on Character Frequency
无
2003-01-01
Based on the study of single pattern matching, MBF algorithm is proposed by imitating the string searching procedure of human. The algorithm preprocesses the pattern by using the idea of Quick Search algorithm and the already-matched pattern psefix and suffix information. In searching phase, the algorithm makes use of the!character using frequency and the continue-skip idea. The experiment shows that MBF algorithm is more efficient than other algorithms.
A Hybrid Algorithm for Satellite Data Transmission Schedule Based on Genetic Algorithm
LI Yun-feng; WU Xiao-yue
2008-01-01
A hybrid scheduling algorithm based on genetic algorithm is proposed in this paper for reconnaissance satellite data transmission. At first, based on description of satellite data transmission request, satellite data transmission task modal and satellite data transmission scheduling problem model are established. Secondly, the conflicts in scheduling are discussed. According to the meaning of possible conflict, the method to divide possible conflict task set is given. Thirdly, a hybrid algorithm which consists of genetic algorithm and heuristic information is presented. The heuristic information comes from two concepts, conflict degree and conflict number. Finally, an example shows the algorithm's feasibility and performance better than other traditional algorithms.
Generalized Rule Induction Based on Immune Algorithm
郑建国; 刘芳; 焦李成
2002-01-01
A generalized rule induction mechanism, immune algorithm, for knowledge bases is building an inheritance hierarchy of classes based on the content of their knowledge objects. This hierarchy facilitates group-related processing tasks such as answering set queries, discriminating between objects, finding similarities among objects, etc. Building this hierarchy is a difficult task for knowledge engineers. Conceptual induction may be used to automate or assist engineers in the creation of such a classification structure. This paper introduces a new conceptual rule induction method, which addresses the problem of clustering large amounts of structured objects. The conditions under which the method is applicable are discussed.
Continuous Attributes Discretization Algorithm based on FPGA
Guoqiang Sun
2013-07-01
Full Text Available The paper addresses the problem of Discretization of continuous attributes in rough set. Discretization of continuous attributes is an important part of rough set theory because most of data that we usually gain are continuous data. In order to improve processing speed of discretization, we propose a FPGA-based discretization algorithm of continuous attributes making use of the speed advantage of FPGA. Combined attributes dependency degree of rough ret, the discretization system was divided into eight modules according to block design. This method can save much time of pretreatment in rough set and improve operation efficiency. Extensive experiments on a certain fighter fault diagnosis validate the effectiveness of the algorithm.
Multi-Agent Reinforcement Learning Algorithm Based on Action Prediction
TONG Liang; LU Ji-lian
2006-01-01
Multi-agent reinforcement learning algorithms are studied. A prediction-based multi-agent reinforcement learning algorithm is presented for multi-robot cooperation task. The multi-robot cooperation experiment based on multi-agent inverted pendulum is made to test the efficency of the new algorithm, and the experiment results show that the new algorithm can achieve the cooperation strategy much faster than the primitive multiagent reinforcement learning algorithm.
Algorithm Research of Individualized Travelling Route Recommendation Based on Similarity
Xue Shan
2015-01-01
Full Text Available Although commercial recommendation system has made certain achievement in travelling route development, the recommendation system is facing a series of challenges because of people’s increasing interest in travelling. It is obvious that the core content of the recommendation system is recommendation algorithm. The advantages of recommendation algorithm can bring great effect to the recommendation system. Based on this, this paper applies traditional collaborative filtering algorithm for analysis. Besides, illustrating the deficiencies of the algorithm, such as the rating unicity and rating matrix sparsity, this paper proposes an improved algorithm combing the multi-similarity algorithm based on user and the element similarity algorithm based on user, so as to compensate for the deficiencies that traditional algorithm has within a controllable range. Experimental results have shown that the improved algorithm has obvious advantages in comparison with the traditional one. The improved algorithm has obvious effect on remedying the rating matrix sparsity and rating unicity.
Asian Option Pricing Based on Genetic Algorithms
YunzhongLiu; HuiyuXuan
2004-01-01
The cross-fertilization between artificial intelligence and computational finance has resulted in some of the most active research areas in financial engineering. One direction is the application of machine learning techniques to pricing financial products, which is certainly one of the most complex issues in finance. In the literature, when the interest rate,the mean rate of return and the volatility of the underlying asset follow general stochastic processes, the exact solution is usually not available. In this paper, we shall illustrate how genetic algorithms (GAs), as a numerical approach, can be potentially helpful in dealing with pricing. In particular, we test the performance of basic genetic algorithms by using it to the determination of prices of Asian options, whose exact solutions is known from Black-Scholesoption pricing theory. The solutions found by basic genetic algorithms are compared with the exact solution, and the performance of GAs is ewluated accordingly. Based on these ewluations, some limitations of GAs in option pricing are examined and possible extensions to future works are also proposed.
Improved pulse laser ranging algorithm based on high speed sampling
Gao, Xuan-yi; Qian, Rui-hai; Zhang, Yan-mei; Li, Huan; Guo, Hai-chao; He, Shi-jie; Guo, Xiao-kang
2016-10-01
Narrow pulse laser ranging achieves long-range target detection using laser pulse with low divergent beams. Pulse laser ranging is widely used in military, industrial, civil, engineering and transportation field. In this paper, an improved narrow pulse laser ranging algorithm is studied based on the high speed sampling. Firstly, theoretical simulation models have been built and analyzed including the laser emission and pulse laser ranging algorithm. An improved pulse ranging algorithm is developed. This new algorithm combines the matched filter algorithm and the constant fraction discrimination (CFD) algorithm. After the algorithm simulation, a laser ranging hardware system is set up to implement the improved algorithm. The laser ranging hardware system includes a laser diode, a laser detector and a high sample rate data logging circuit. Subsequently, using Verilog HDL language, the improved algorithm is implemented in the FPGA chip based on fusion of the matched filter algorithm and the CFD algorithm. Finally, the laser ranging experiment is carried out to test the improved algorithm ranging performance comparing to the matched filter algorithm and the CFD algorithm using the laser ranging hardware system. The test analysis result demonstrates that the laser ranging hardware system realized the high speed processing and high speed sampling data transmission. The algorithm analysis result presents that the improved algorithm achieves 0.3m distance ranging precision. The improved algorithm analysis result meets the expected effect, which is consistent with the theoretical simulation.
A new optimization algorithm based on chaos
无
2006-01-01
In this article, some methods are proposed for enhancing the converging velocity of the COA (chaos optimization algorithm) based on using carrier wave two times, which can greatly increase the speed and efficiency of the first carrier wave's search for the optimal point in implementing the sophisticated searching during the second carrier wave is faster and more accurate.In addition, the concept of using the carrier wave three times is proposed and put into practice to tackle the multi-variables optimization problems, where the searching for the optimal point of the last several variables is frequently worse than the first several ones.
Function Optimization Based on Quantum Genetic Algorithm
Ying Sun; Hegen Xiong
2014-01-01
Optimization method is important in engineering design and application. Quantum genetic algorithm has the characteristics of good population diversity, rapid convergence and good global search capability and so on. It combines quantum algorithm with genetic algorithm. A novel quantum genetic algorithm is proposed, which is called Variable-boundary-coded Quantum Genetic Algorithm (vbQGA) in which qubit chromosomes are collapsed into variable-boundary-coded chromosomes instead of binary-coded c...
Function Optimization Based on Quantum Genetic Algorithm
Ying Sun; Yuesheng Gu; Hegen Xiong
2013-01-01
Quantum genetic algorithm has the characteristics of good population diversity, rapid convergence and good global search capability and so on.It combines quantum algorithm with genetic algorithm. A novel quantum genetic algorithm is proposed ,which is called variable-boundary-coded quantum genetic algorithm (vbQGA) in which qubit chromosomes are collapsed into variableboundary- coded chromosomes instead of binary-coded chromosomes. Therefore much shorter chromosome strings can be gained.The m...
Dynamic route guidance algorithm based algorithm based on artificial immune system
无
2007-01-01
To improve the performance of the K-shortest paths search in intelligent traffic guidance systems,this paper proposes an optimal search algorithm based on the intelligent optimization search theory and the memphor mechanism of vertebrate immune systems.This algorithm,applied to the urban traffic network model established by the node-expanding method,can expediently realize K-shortest paths search in the urban traffic guidance systems.Because of the immune memory and global parallel search ability from artificial immune systems,K shortest paths can be found without any repeat,which indicates evidently the superiority of the algorithm to the conventional ones.Not only does it perform a better parallelism,the algorithm also prevents premature phenomenon that often occurs in genetic algorithms.Thus,it is especially suitable for real-time requirement of the traffic guidance system and other engineering optimal applications.A case study verifies the efficiency and the practicability of the algorithm aforementioned.
Cognitive radio resource allocation based on coupled chaotic genetic algorithm
Zu Yun-Xiao; Zhou Jie; Zeng Chang-Chang
2010-01-01
A coupled chaotic genetic algorithm for cognitive radio resource allocation which is based on genetic algorithm and coupled Logistic map is proposed. A fitness function for cognitive radio resource allocation is provided. Simulations are conducted for cognitive radio resource allocation by using the coupled chaotic genetic algorithm, simple genetic algorithm and dynamic allocation algorithm respectively. The simulation results show that, compared with simple genetic and dynamic allocation algorithm, coupled chaotic genetic algorithm reduces the total transmission power and bit error rate in cognitive radio system, and has faster convergence speed.
Cognitive radio resource allocation based on coupled chaotic genetic algorithm
Zu, Yun-Xiao; Zhou, Jie; Zeng, Chang-Chang
2010-11-01
A coupled chaotic genetic algorithm for cognitive radio resource allocation which is based on genetic algorithm and coupled Logistic map is proposed. A fitness function for cognitive radio resource allocation is provided. Simulations are conducted for cognitive radio resource allocation by using the coupled chaotic genetic algorithm, simple genetic algorithm and dynamic allocation algorithm respectively. The simulation results show that, compared with simple genetic and dynamic allocation algorithm, coupled chaotic genetic algorithm reduces the total transmission power and bit error rate in cognitive radio system, and has faster convergence speed.
An assembly sequence planning method based on composite algorithm
Enfu LIU
2016-02-01
Full Text Available To solve the combination explosion problem and the blind searching problem in assembly sequence planning of complex products, an assembly sequence planning method based on composite algorithm is proposed. In the composite algorithm, a sufficient number of feasible assembly sequences are generated using formalization reasoning algorithm as the initial population of genetic algorithm. Then fuzzy knowledge of assembly is integrated into the planning process of genetic algorithm and ant algorithm to get the accurate solution. At last, an example is conducted to verify the feasibility of composite algorithm.
A Survey of Grid Based Clustering Algorithms
MR ILANGO
2010-08-01
Full Text Available Cluster Analysis, an automatic process to find similar objects from a database, is a fundamental operation in data mining. A cluster is a collection of data objects that are similar to one another within the same cluster and are dissimilar to the objects in other clusters. Clustering techniques have been discussed extensively in SimilaritySearch, Segmentation, Statistics, Machine Learning, Trend Analysis, Pattern Recognition and Classification [1]. Clustering methods can be classified into i Partitioning methods ii Hierarchical methods iii Density-based methods iv Grid-based methods v Model-based methods. Grid based methods quantize the object space into a finite number of cells (hyper-rectangles and then perform the required operations on the quantized space. The main advantage of Grid based method is its fast processing time which depends on number of cells in each dimension in quantized space. In this research paper, we present some of the grid based methods such as CLIQUE (CLustering In QUEst [2], STING (STatistical INformation Grid [3], MAFIA (Merging of Adaptive Intervals Approach to Spatial Data Mining [4], Wave Cluster [5]and O-CLUSTER (Orthogonal partitioning CLUSTERing [6], as a survey andalso compare their effectiveness in clustering data objects. We also present some of the latest developments in Grid Based methods such as Axis Shifted Grid Clustering Algorithm [7] and Adaptive Mesh Refinement [Wei-Keng Liao etc] [8] to improve the processing time of objects.
ALGORITHM FOR GENERATING DEM BASED ON CONE
无
2000-01-01
Digital elevation model (DEM) has a variety of applications in GIS and CAD.It is the basic model for generating three-dimensional terrain feature.Generally speaking,there are two methods for building DEM.One is based upon the digital terrain model of discrete points,and is characterized by fast speed and low precision.The other is based upon triangular digital terrain model,and slow speed and high precision are the features of the method.Combining the advantages of the two methods,an algorithm for generating DEM with discrete points is presented in this paper.When interpolating elevation,this method can create a triangle which includes interpolating point and the elevation of the interpolating point can be obtained from the triangle.The method has the advantage of fast speed,high precision and less memory.
A Genetic Algorithm-Based Feature Selection
Babatunde Oluleye
2014-07-01
Full Text Available This article details the exploration and application of Genetic Algorithm (GA for feature selection. Particularly a binary GA was used for dimensionality reduction to enhance the performance of the concerned classifiers. In this work, hundred (100 features were extracted from set of images found in the Flavia dataset (a publicly available dataset. The extracted features are Zernike Moments (ZM, Fourier Descriptors (FD, Lengendre Moments (LM, Hu 7 Moments (Hu7M, Texture Properties (TP and Geometrical Properties (GP. The main contributions of this article are (1 detailed documentation of the GA Toolbox in MATLAB and (2 the development of a GA-based feature selector using a novel fitness function (kNN-based classification error which enabled the GA to obtain a combinatorial set of feature giving rise to optimal accuracy. The results obtained were compared with various feature selectors from WEKA software and obtained better results in many ways than WEKA feature selectors in terms of classification accuracy
A Trust-region-based Sequential Quadratic Programming Algorithm
Henriksen, Lars Christian; Poulsen, Niels Kjølstad
This technical note documents the trust-region-based sequential quadratic programming algorithm used in other works by the authors. The algorithm seeks to minimize a convex nonlinear cost function subject to linear inequalty constraints and nonlinear equality constraints....
Speech Enhancement based on Compressive Sensing Algorithm
Sulong, Amart; Gunawan, Teddy S.; Khalifa, Othman O.; Chebil, Jalel
2013-12-01
There are various methods, in performance of speech enhancement, have been proposed over the years. The accurate method for the speech enhancement design mainly focuses on quality and intelligibility. The method proposed with high performance level. A novel speech enhancement by using compressive sensing (CS) is a new paradigm of acquiring signals, fundamentally different from uniform rate digitization followed by compression, often used for transmission or storage. Using CS can reduce the number of degrees of freedom of a sparse/compressible signal by permitting only certain configurations of the large and zero/small coefficients, and structured sparsity models. Therefore, CS is significantly provides a way of reconstructing a compressed version of the speech in the original signal by taking only a small amount of linear and non-adaptive measurement. The performance of overall algorithms will be evaluated based on the speech quality by optimise using informal listening test and Perceptual Evaluation of Speech Quality (PESQ). Experimental results show that the CS algorithm perform very well in a wide range of speech test and being significantly given good performance for speech enhancement method with better noise suppression ability over conventional approaches without obvious degradation of speech quality.
PDE Based Algorithms for Smooth Watersheds.
Hodneland, Erlend; Tai, Xue-Cheng; Kalisch, Henrik
2016-04-01
Watershed segmentation is useful for a number of image segmentation problems with a wide range of practical applications. Traditionally, the tracking of the immersion front is done by applying a fast sorting algorithm. In this work, we explore a continuous approach based on a geometric description of the immersion front which gives rise to a partial differential equation. The main advantage of using a partial differential equation to track the immersion front is that the method becomes versatile and may easily be stabilized by introducing regularization terms. Coupling the geometric approach with a proper "merging strategy" creates a robust algorithm which minimizes over- and under-segmentation even without predefined markers. Since reliable markers defined prior to segmentation can be difficult to construct automatically for various reasons, being able to treat marker-free situations is a major advantage of the proposed method over earlier watershed formulations. The motivation for the methods developed in this paper is taken from high-throughput screening of cells. A fully automated segmentation of single cells enables the extraction of cell properties from large data sets, which can provide substantial insight into a biological model system. Applying smoothing to the boundaries can improve the accuracy in many image analysis tasks requiring a precise delineation of the plasma membrane of the cell. The proposed segmentation method is applied to real images containing fluorescently labeled cells, and the experimental results show that our implementation is robust and reliable for a variety of challenging segmentation tasks.
A Text Categorization Algorithm Based on Sense Group
Jing Wan
2013-02-01
Full Text Available Giving further consideration on linguistic feature, this study proposes an algorithm of Chinese text categorization based on sense group. The algorithm extracts sense group by analyzing syntactic and semantic properties of Chinese texts and builds the category sense group library. SVM is used for the experiment of text categorization. The experimental results show that the precision and recall of the new algorithm based on sense group is better than that of traditional algorithms.
POWER OPTIMIZATION ALGORITHM BASED ON XNOR/OR LOGIC
Wang Pengjun; Lu Jingang; Xu Jian; Dai Jing
2009-01-01
Based on the investigation of the XNOR/OR logical expression and the propagation algorithm of signal probability, a low power synthesis algorithm based on the XNOR/OR logic is proposed in this paper. The proposed algorithm has been implemented with C language. Fourteen Microelectronics Center North Carolina (MCNC) benchmarks are tested, and the results show that the proposed algorithm not only significantly reduces the average power consumption up to 27% without area and delay compensations, but also makes the runtime shorter.
Performance evaluation of sensor allocation algorithm based on covariance control
无
2005-01-01
The covariance control capability of sensor allocation algorithms based on covariance control strategy is an important index to evaluate the performance of these algorithms. Owing to lack of standard performance metric indices to evaluate covariance control capability, sensor allocation ratio, etc, there are no guides to follow in the design procedure of sensor allocation algorithm in practical applications. To meet these demands, three quantified performance metric indices are presented, which are average covariance misadjustment quantity (ACMQ), average sensor allocation ratio (ASAR) and matrix metric influence factor (MMIF), where ACMQ, ASAR and MMIF quantify the covariance control capability, the usage of sensor resources and the robustness of sensor allocation algorithm, respectively. Meanwhile, a covariance adaptive sensor allocation algorithm based on a new objective function is proposed to improve the covariance control capability of the algorithm based on information gain. The experiment results show that the proposed algorithm have the advantage over the preceding sensor allocation algorithm in covariance control capability and robustness.
Chaos-Based Multipurpose Image Watermarking Algorithm
ZHU Congxu; LIAO Xuefeng; LI Zhihua
2006-01-01
To achieve the goal of image content authentication and copyright protection simultaneously, this paper presents a novel image dual watermarking method based on chaotic map. Firstly, the host image was split into many nonoverlapping small blocks, and the block-wise discrete cosine transform (DCT) is computed. Secondly, the robust watermarks, shuffled by the chaotic sequences, are embedded in the DC coefficients of blocks to achieve the goal of copyright protection. The semi-fragile watermarks, generated by chaotic map, are embedded in the AC coefficients of blocks to obtain the aim of image authentication. Both of them can be extracted without the original image. Simulation results demonstrate the effectiveness of our algorithm in terms of robustness and fragility.
Review: Image Encryption Using Chaos Based algorithms
Er. Ankita Gaur
2014-03-01
Full Text Available Due to the development in the field of network technology and multimedia applications, every minute thousands of messages which can be text, images, audios, videos are created and transmitted over wireless network. Improper delivery of the message may leads to the leakage of important information. So encryption is used to provide security. In last few years, variety of image encryption algorithms based on chaotic system has been proposed to protect image from unauthorized access. 1-D chaotic system using logistic maps has weak security, small key space and due to the floating of pixel values, some data lose occurs and proper decryption of image becomes impossible. In this paper different chaotic maps such as Arnold cat map, sine map, logistic map, tent map have been studied.
An intersection algorithm based on transformation
CHEN Xiao-xia; YONG Jun-hai; CHEN Yu-jian
2006-01-01
How to obtain intersection of curves and surfaces is a fundamental problem in many areas such as computer graphics,CAD/CAM,computer animation,and robotics.Especially,how to deal with singular cases,such as tangency or superposition,is a key problem in obtaining intersection results.A method for solving the intersection problem based on the coordinate transformation is presented.With the Lagrange multiplier method,the minimum distance between the center of a circle and a quadric surface is given as well.Experience shows that the coordinate transformation could significantly simplify the method for calculating intersection to the tangency condition.It can improve the stability of the intersection of given curves and surfaces in singularity cases.The new algorithm is applied in a three dimensional CAD software (GEMS),produced by Tsinghua University.
An Improved Particle Swarm Optimization Algorithm Based on Ensemble Technique
SHI Yan; HUANG Cong-ming
2006-01-01
An improved particle swarm optimization (PSO) algorithm based on ensemble technique is presented. The algorithm combines some previous best positions (pbest) of the particles to get an ensemble position (Epbest), which is used to replace the global best position (gbest). It is compared with the standard PSO algorithm invented by Kennedy and Eberhart and some improved PSO algorithms based on three different benchmark functions. The simulation results show that the improved PSO based on ensemble technique can get better solutions than the standard PSO and some other improved algorithms under all test cases.
A New Aloha Anti-Collision Algorithm Based on CDMA
Bai, Enjian; Feng, Zhu
The tags' collision is a common problem in RFID (radio frequency identification) system. The problem has affected the integrity of the data transmission during the process of communication in the RFID system. Based on analysis of the existing anti-collision algorithm, a novel anti-collision algorithm is presented. The new algorithm combines the group dynamic frame slotted Aloha algorithm with code division multiple access technology. The algorithm can effectively reduce the collision probability between tags. Under the same number of tags, the algorithm is effective in reducing the reader recognition time and improve overall system throughput rate.
A research on fast FCM algorithm based on weighted sample
KUANG Ping; ZHU Qing-xin; WANG Ming-wen; CHEN Xu-dong; QING Li
2006-01-01
To improve the computational performance of the fuzzy C-means (FCM) algorithm used in dataset clustering with large numbers,the concepts of the equivalent samples and the weighting samples based on eigenvalue distribution of the samples in the feature space were introduced and a novel fast cluster algorithm named weighted fuzzy C-means (WFCM) algorithm was put forward,which came from the traditional FCM algorithm.It was proved that the duster results were equivalent in dataset with two different cluster algorithms:WFCM and FCM.Furthermore,the WFCM algorithm had better computational performance than the ordinary FCM algorithm.The experiment of the gray image segmentation showed that the WFCM algorithm is a fast and effective cluster algorithm.
An improved localization algorithm based on genetic algorithm in wireless sensor networks.
Peng, Bo; Li, Lei
2015-04-01
Wireless sensor network (WSN) are widely used in many applications. A WSN is a wireless decentralized structure network comprised of nodes, which autonomously set up a network. The node localization that is to be aware of position of the node in the network is an essential part of many sensor network operations and applications. The existing localization algorithms can be classified into two categories: range-based and range-free. The range-based localization algorithm has requirements on hardware, thus is expensive to be implemented in practice. The range-free localization algorithm reduces the hardware cost. Because of the hardware limitations of WSN devices, solutions in range-free localization are being pursued as a cost-effective alternative to more expensive range-based approaches. However, these techniques usually have higher localization error compared to the range-based algorithms. DV-Hop is a typical range-free localization algorithm utilizing hop-distance estimation. In this paper, we propose an improved DV-Hop algorithm based on genetic algorithm. Simulation results show that our proposed algorithm improves the localization accuracy compared with previous algorithms.
Uzawa Type Algorithm Based on Dual Mixed Variational Formulation
王光辉; 王烈衡
2002-01-01
Based on the dual mixed variational formulation with three variants (stress,displacement, displacement on contact boundary ) and the unilateral beaming problem of finite element discretization, an Uzawa type iterative algorithm is presented. The convergence of this iterative algorithm is proved, and then the efficiency of the algorithm is tested by a numerical example.
Replication-based Inference Algorithms for Hard Computational Problems
Alamino, Roberto C.; Neirotti, Juan P.; Saad, David
2013-01-01
Inference algorithms based on evolving interactions between replicated solutions are introduced and analyzed on a prototypical NP-hard problem - the capacity of the binary Ising perceptron. The efficiency of the algorithm is examined numerically against that of the parallel tempering algorithm, showing improved performance in terms of the results obtained, computing requirements and simplicity of implementation.
Network Intrusion Detection based on GMKL Algorithm
Li Yuxiang
2013-06-01
Full Text Available According to the 31th statistical reports of China Internet network information center (CNNIC, by the end of December 2012, the number of Chinese netizens has reached 564 million, and the scale of mobile Internet users also reached 420 million. But when the network brings great convenience to people's life, it also brings huge threat in the life of people. So through collecting and analyzing the information in the computer system or network we can detect any possible behaviors that can damage the availability, integrity and confidentiality of the computer resource, and make timely treatment to these behaviors which have important research significance to improve the operation environment of network and network service. At present, the Neural Network, Support Vector machine (SVM and Hidden Markov Model, Fuzzy inference and Genetic Algorithms are introduced into the research of network intrusion detection, trying to build a healthy and secure network operation environment. But most of these algorithms are based on the total sample and it also hypothesizes that the number of the sample is infinity. But in the field of network intrusion the collected data often cannot meet the above requirements. It often shows high latitudes, variability and small sample characteristics. For these data using traditional machine learning methods are hard to get ideal results. In view of this, this paper proposed a Generalized Multi-Kernel Learning method to applied to network intrusion detection. The Generalized Multi-Kernel Learning method can be well applied to large scale sample data, dimension complex, containing a large number of heterogeneous information and so on. The experimental results show that applying GMKL to network attack detection has high classification precision and low abnormal practical precision.
An incremental clustering algorithm based on Mahalanobis distance
Aik, Lim Eng; Choon, Tan Wee
2014-12-01
Classical fuzzy c-means clustering algorithm is insufficient to cluster non-spherical or elliptical distributed datasets. The paper replaces classical fuzzy c-means clustering euclidean distance with Mahalanobis distance. It applies Mahalanobis distance to incremental learning for its merits. A Mahalanobis distance based fuzzy incremental clustering learning algorithm is proposed. Experimental results show the algorithm is an effective remedy for the defect in fuzzy c-means algorithm but also increase training accuracy.
Saudi License Plate Recognition Algorithm Based on Support Vector Machine
Khaled Suwais; Rana Al-Otaibi; Ali Alshahrani
2013-01-01
License plate recognition (LPR) is an image processing technology that is used to identify vehicles by their license plates. This paper presents a license plate recognition algorithm for Saudi car plates based on the support vector machine (SVM) algorithm. The new algorithm is efficient in recognizing the vehicles from the Arabic part of the plate. The performance of the system has been investigated and analyzed. The recognition accuracy of the algorithm is about 93.3%.
A new classification algorithm based on RGH-tree search
无
2007-01-01
In this paper, we put forward a new classification algorithm based on RGH-Tree search and perform the classification analysis and comparison study. This algorithm can save computing resource and increase the classification efficiency. The experiment shows that this algorithm can get better effect in dealing with three dimensional multi-kind data. We find that the algorithm has better generalization ability for small training set and big testing result.
The Result Integration Algorithm Based on Matching Strategy
无
2006-01-01
The following paper provides a new algorithm: a result integration algorithm based on matching strategy. The algorithm extracts the title and the abstract of Web pages, calculates the relevance between the query string and the Web pages, decides the Web pages accepted, rejected and sorts them out in user interfaces. The experiment results indicate obviously that the new algorithms improve the precision of meta-search engine. This technique is very useful to meta-search engine.
An Incremental Algorithm of Text Clustering Based on Semantic Sequences
FENG Zhonghui; SHEN Junyi; BAO Junpeng
2006-01-01
This paper proposed an incremental textclustering algorithm based on semantic sequence.Using similarity relation of semantic sequences and calculating the cover of similarity semantic sequences set, the candidate cluster with minimum entropy overlap value was selected as a result cluster every time in this algorithm.The comparison of experimental results shows that the precision of the algorithm is higher than other algorithms under same conditions and this is obvious especially on long documents set.
A generalized GPU-based connected component labeling algorithm
Komura, Yukihiro
2016-01-01
We propose a generalized GPU-based connected component labeling (CCL) algorithm that can be applied to both various lattices and to non-lattice environments in a uniform fashion. We extend our recent GPU-based CCL algorithm without the use of conventional iteration to the generalized method. As an application of this algorithm, we deal with the bond percolation problem. We investigate bond percolation on the honeycomb and triangle lattices to confirm the correctness of this algorithm. Moreover, we deal with bond percolation on the Bethe lattice as a substitute for a network structure, and demonstrate the performance of this algorithm on those lattices.
Fixed-point blind source separation algorithm based on ICA
Hongyan LI; Jianfen MA; Deng'ao LI; Huakui WANG
2008-01-01
This paper introduces the fixed-point learning algorithm based on independent component analysis (ICA);the model and process of this algorithm and simulation results are presented.Kurtosis was adopted as the estimation rule of independence.The results of the experiment show that compared with the traditional ICA algorithm based on random grads,this algorithm has advantages such as fast convergence and no necessity for any dynamic parameter,etc.The algorithm is a highly efficient and reliable method in blind signal separation.
Adaptive Central Force Optimization Algorithm Based on the Stability Analysis
Weiyi Qian
2015-01-01
Full Text Available In order to enhance the convergence capability of the central force optimization (CFO algorithm, an adaptive central force optimization (ACFO algorithm is presented by introducing an adaptive weight and defining an adaptive gravitational constant. The adaptive weight and gravitational constant are selected based on the stability theory of discrete time-varying dynamic systems. The convergence capability of ACFO algorithm is compared with the other improved CFO algorithm and evolutionary-based algorithm using 23 unimodal and multimodal benchmark functions. Experiments results show that ACFO substantially enhances the performance of CFO in terms of global optimality and solution accuracy.
Clonal Selection Based Memetic Algorithm for Job Shop Scheduling Problems
Jin-hui Yang; Liang Sun; Heow Pueh Lee; Yun Qian; Yan-chun Liang
2008-01-01
A clonal selection based memetic algorithm is proposed for solving job shop scheduling problems in this paper. In the proposed algorithm, the clonal selection and the local search mechanism are designed to enhance exploration and exploitation. In the clonal selection mechanism, clonal selection, hypermutation and receptor edit theories are presented to construct an evolutionary searching mechanism which is used for exploration. In the local search mechanism, a simulated annealing local search algorithm based on Nowicki and Smutnicki's neighborhood is presented to exploit local optima. The proposed algorithm is examined using some well-known benchmark problems. Numerical results validate the effectiveness of the proposed algorithm.
Adaptive RED algorithm based on minority game
Wei, Jiaolong; Lei, Ling; Qian, Jingjing
2007-11-01
With more and more applications appearing and the technology developing in the Internet, only relying on terminal system can not satisfy the complicated demand of QoS network. Router mechanisms must be participated into protecting responsive flows from the non-responsive. Routers mainly use active queue management mechanism (AQM) to avoid congestion. In the point of interaction between the routers, the paper applies minority game to describe the interaction of the users and observes the affection on the length of average queue. The parameters α, β of ARED being hard to confirm, adaptive RED based on minority game can depict the interactions of main body and amend the parameter α, β of ARED to the best. Adaptive RED based on minority game optimizes ARED and realizes the smoothness of average queue length. At the same time, this paper extends the network simulator plat - NS by adding new elements. Simulation has been implemented and the results show that new algorithm can reach the anticipative objects.
Web Based Genetic Algorithm Using Data Mining
Ashiqur Rahman
2016-09-01
Full Text Available This paper presents an approach for classifying students in order to predict their final grade based on features extracted from logged data in an education web-based system. A combination of multiple classifiers leads to a significant improvement in classification performance. Through weighting the feature vectors using a Genetic Algorithm we can optimize the prediction accuracy and get a marked improvement over raw classification. It further shows that when the number of features is few; feature weighting is works better than just feature selection. Many leading educational institutions are working to establish an online teaching and learning presence. Several systems with different capabilities and approaches have been developed to deliver online education in an academic setting. In particular, Michigan State University (MSU has pioneered some of these systems to provide an infrastructure for online instruction. The research presented here was performed on a part of the latest online educational system developed at MSU, the Learning Online Network with Computer-Assisted Personalized Approach (LON-CAPA
DYNAMIC LABELING BASED FPGA DELAY OPTIMIZATION ALGORITHM
吕宗伟; 林争辉; 张镭
2001-01-01
DAG-MAP is an FPGA technology mapping algorithm for delay optimization and the labeling phase is the algorithm's kernel. This paper studied the labeling phase and presented an improved labeling method. It is shown through the experimental results on MCNC benchmarks that the improved method is more effective than the original method while the computation time is almost the same.
ADAPTIVE FUSION ALGORITHMS BASED ON WEIGHTED LEAST SQUARE METHOD
SONG Kaichen; NIE Xili
2006-01-01
Weighted fusion algorithms, which can be applied in the area of multi-sensor data fusion,are advanced based on weighted least square method. A weighted fusion algorithm, in which the relationship between weight coefficients and measurement noise is established, is proposed by giving attention to the correlation of measurement noise. Then a simplified weighted fusion algorithm is deduced on the assumption that measurement noise is uncorrelated. In addition, an algorithm, which can adjust the weight coefficients in the simplified algorithm by making estimations of measurement noise from measurements, is presented. It is proved by emulation and experiment that the precision performance of the multi-sensor system based on these algorithms is better than that of the multi-sensor system based on other algorithms.
Gradient-based Taxis Algorithms for Network Robotics
Blum, Christian; Hafner, Verena V.
2014-01-01
Finding the physical location of a specific network node is a prototypical task for navigation inside a wireless network. In this paper, we consider in depth the implications of wireless communication as a measurement input of gradient-based taxis algorithms. We discuss how gradients can be measured and determine the errors of this estimation. We then introduce a gradient-based taxis algorithm as an example of a family of gradient-based, convergent algorithms and discuss its convergence in th...
LEACH Algorithm Based on Load Balancing
Wangang Wang
2013-09-01
Full Text Available This paper discusses advantages of LEACH Algorithm and the existing improved model which takes the famous hierarchy clustering routing protocol LEACH Algorithm as researching object. Then the paper indicates the problem that in the algorithm capacity factor of cluster head node is not taken into account leading the structure of clusters to be not so reasonable. This research discusses an energy-uniform cluster and cluster head selecting mechanism in which “Pseudo cluster head” concept is introduced in order to coordinate with “Load Monitor” Mechanism and “Load Leisure” Mechanism to maintain load balancing of cluster head character and stability of network topology. On the basis of LEACH Protocol improving algorithm of LEACH-C, CEFL and DCHS. NS2 simulation instrument is applied to do simulation analysis on the improved algorithm. Simulation result shows that LEACH-P Protocol effectively increase energy utilization efficiency, lengthens network lifetime and balances network load.
Combined string searching algorithm based on knuth-morris- pratt and boyer-moore algorithms
Tsarev, R. Yu; Chernigovskiy, A. S.; Tsareva, E. A.; Brezitskaya, V. V.; Nikiforov, A. Yu; Smirnov, N. A.
2016-04-01
The string searching task can be classified as a classic information processing task. Users either encounter the solution of this task while working with text processors or browsers, employing standard built-in tools, or this task is solved unseen by the users, while they are working with various computer programmes. Nowadays there are many algorithms for solving the string searching problem. The main criterion of these algorithms’ effectiveness is searching speed. The larger the shift of the pattern relative to the string in case of pattern and string characters’ mismatch is, the higher is the algorithm running speed. This article offers a combined algorithm, which has been developed on the basis of well-known Knuth-Morris-Pratt and Boyer-Moore string searching algorithms. These algorithms are based on two different basic principles of pattern matching. Knuth-Morris-Pratt algorithm is based upon forward pattern matching and Boyer-Moore is based upon backward pattern matching. Having united these two algorithms, the combined algorithm allows acquiring the larger shift in case of pattern and string characters’ mismatch. The article provides an example, which illustrates the results of Boyer-Moore and Knuth-Morris- Pratt algorithms and combined algorithm’s work and shows advantage of the latter in solving string searching problem.
An Innovative Thinking-Based Intelligent Information Fusion Algorithm
Huimin Lu
2013-01-01
Full Text Available This study proposes an intelligent algorithm that can realize information fusion in reference to the relative research achievements in brain cognitive theory and innovative computation. This algorithm treats knowledge as core and information fusion as a knowledge-based innovative thinking process. Furthermore, the five key parts of this algorithm including information sense and perception, memory storage, divergent thinking, convergent thinking, and evaluation system are simulated and modeled. This algorithm fully develops innovative thinking skills of knowledge in information fusion and is a try to converse the abstract conception of brain cognitive science to specific and operable research routes and strategies. Furthermore, the influences of each parameter of this algorithm on algorithm performance are analyzed and compared with those of classical intelligent algorithms trough test. Test results suggest that the algorithm proposed in this study can obtain the optimum problem solution by less target evaluation times, improve optimization effectiveness, and achieve the effective fusion of information.
Jiang Ting
2010-01-01
Full Text Available We optimize the cluster structure to solve problems such as the uneven energy consumption of the radar sensor nodes and random cluster head selection in the traditional clustering routing algorithm. According to the defined cost function for clusters, we present the clustering algorithm which is based on radio-free space path loss. In addition, we propose the energy and distance pheromones based on the residual energy and aggregation of the radar sensor nodes. According to bionic heuristic algorithm, a new ant colony-based clustering algorithm for radar sensor networks is also proposed. Simulation results show that this algorithm can get a better balance of the energy consumption and then remarkably prolong the lifetime of the radar sensor network.
曾宪钊; 成冀; 安欣; 方礼明
2002-01-01
This paper introduces a new Air Combat Intelligence Simulation System (ACISS) in a 32 versus 32 air combat, describes three methods: Genetic Algorithms (GA) in the multi-targeting decision and Evading Missile Rule Base learning, Neural Networks (NN) in the maneuvering decision, and Time Effectiveness Algorithm (TEA) in the adjudicating an air combat and the evaluating evading missile effectiveness.
Parallel Implementation of Classification Algorithms Based on Cloud Computing Environment
Wenbo Wang
2012-09-01
Full Text Available As an important task of data mining, Classification has been received considerable attention in many applications, such as information retrieval, web searching, etc. The enlarging volumes of information emerging by the progress of technology and the growing individual needs of data mining, makes classifying of very large scale of data a challenging task. In order to deal with the problem, many researchers try to design efficient parallel classification algorithms. This paper introduces the classification algorithms and cloud computing briefly, based on it analyses the bad points of the present parallel classification algorithms, then addresses a new model of parallel classifying algorithms. And it mainly introduces a parallel Naïve Bayes classification algorithm based on MapReduce, which is a simple yet powerful parallel programming technique. The experimental results demonstrate that the proposed algorithm improves the original algorithm performance, and it can process large datasets efficiently on commodity hardware.
SAR Image Segmentation Based On Hybrid PSOGSA Optimization Algorithm
Amandeep Kaur
2014-09-01
Full Text Available Image segmentation is useful in many applications. It can identify the regions of interest in a scene or annotate the data. It categorizes the existing segmentation algorithm into region-based segmentation, data clustering, and edge-base segmentation. Region-based segmentation includes the seeded and unseeded region growing algorithms, the JSEG, and the fast scanning algorithm. Due to the presence of speckle noise, segmentation of Synthetic Aperture Radar (SAR images is still a challenging problem. We proposed a fast SAR image segmentation method based on Particle Swarm Optimization-Gravitational Search Algorithm (PSO-GSA. In this method, threshold estimation is regarded as a search procedure that examinations for an appropriate value in a continuous grayscale interval. Hence, PSO-GSA algorithm is familiarized to search for the optimal threshold. Experimental results indicate that our method is superior to GA based, AFS based and ABC based methods in terms of segmentation accuracy, segmentation time, and Thresholding.
New Iris Localization Method Based on Chaos Genetic Algorithm
Jia Dongli; Muhammad Khurram Khan; Zhang Jiashu
2005-01-01
This paper present a new method based on Chaos Genetic Algorithm (CGA) to localize the human iris in a given image. First, the iris image is preprocessed to estimate the range of the iris localization, and then CGA is used to extract the boundary of the iris. Simulation results show that the proposed algorithms is efficient and robust, and can achieve sub pixel precision. Because Genetic Algorithms (GAs) can search in a large space, the algorithm does not need accurate estimation of iris center for subsequent localization, and hence can lower the requirement for original iris image processing. On this point, the present localization algirithm is superior to Daugmans algorithm.
A Wire-speed Routing Lookup Algorithm Based on TCAM
李小勇; 王志恒; 白英彩; 刘刚
2004-01-01
An internal structure of Ternary Content Addressable Memory (TCAM) is designed and a Sorting Prefix Block (SPB) algorithm is presented, which is a wire-speed routing lookup algorithm based on TCAM. SPB algorithm makes use of the parallelism of TCAM adequately, and improves the utilization of TCAM by optimum partitions. With the aid of effective management algorithm and memory image, SPB separates critical searching from assistant searching, and improves the searching effect. One performance test indicates that this algorithm can work with different TCAM to meet the requirement of wire-speed routing lookup.
Information criterion based fast PCA adaptive algorithm
Li Jiawen; Li Congxin
2007-01-01
The novel information criterion (NIC) algorithm can find the principal subspace quickly, but it is not an actual principal component analysis (PCA) algorithm and hence it cannot find the orthonormal eigen-space which corresponds to the principal component of input vector.This defect limits its application in practice.By weighting the neural network's output of NIC, a modified novel information criterion (MNIC) algorithm is presented.MNIC extractes the principal components and corresponding eigenvectors in a parallel online learning program, and overcomes the NIC's defect.It is proved to have a single global optimum and nonquadratic convergence rate, which is superior to the conventional PCA online algorithms such as Oja and LMSER.The relationship among Oja, LMSER and MNIC is exhibited.Simulations show that MNIC could converge to the optimum fast.The validity of MNIC is proved.
A Multi-Scale Gradient Algorithm Based on Morphological Operators
无
2000-01-01
Watershed transformation is a powerful morphological tool for image segmentation. However, the performance of the image segmentation methods based on watershed transformation depends largely on the algorithm for computing the gradient of the image to be segmented. In this paper, we present a multi-scale gradient algorithm based on morphological operators for watershed-based image segmentation, with effective handling of both step and blurred edges. We also present an algorithm to eliminate the local minima produced by noise and quantization errors. Experimental results indicate that watershed transformation with the algorithms proposed in this paper produces meaningful segmentations, even without a region-merging step.
Cycle-Based Algorithm Used to Accelerate VHDL Simulation
杨勋; 刘明业
2000-01-01
Cycle-based algorithm has very high performance for the simula-tion of synchronous design, but it is confined to synchronous design and it is not as accurate as event-driven algorithm. In this paper, a revised cycle-based algorithm is proposed and implemented in VHDL simulator. Event-driven simulation engine and cycle-based simulation engine have been imbedded in the same simulation environ-ment and can be used to asynchronous design and synchronous design respectively. Thus the simulation performance is improved without losing the flexibility and ac-curacy of event-driven algorithm.
QOS-BASED MULTICAST ROUTING OPTIMIZATION ALGORITHMS FOR INTERNET
无
2006-01-01
Most of the multimedia applications require strict Quality-of-Service (QoS) guarantee during the communication between a single source and multiple destinations. The paper mainly presents a QoS Multicast Routing algorithms based on Genetic Algorithm (QMRGA). Simulation results demonstrate that the algorithm is capable of discovering a set of QoS-based near optimized, non-dominated multicast routes within a few iterations, even for the networks environment with uncertain parameters.
Improved FCLSD algorithm based on LTE/LTE-A system
Kewen Liu
2011-08-01
Full Text Available In order to meet the high data rate, large capacity and low latency in LTE, advanced MIMO technology has been introduced in LTE system, which becomes one of the core technologies in physical layer. In a variety of MIMO detection algorithms, the ZF and MMSE linear detection algorithms are the most simple, but the performance is poor. MLD algorithm can achieve optimal detection performance, but it’s too complexity to be applied in practice. CLSD algorithm has similar detection performance and lower complexity with the MLD algorithm, but the uncertainty of complexity will bring hardware difficulties. FCLSD algorithm can maximize the advantages of CLSD algorithm and solve difficult problems in practice. Based on advanced FCLSD algorithm and combined with LTE / LTE-A practical system applications, this article designed two improved algorithms. The two improved algorithms can be flexibly and adaptively used in various antenna configurations and modulation scene in LTE / LTE-A spatial multiplexing MIMO system. The Simulation results show that the improved algorithm can achieve an approximate performance to the original FCLSD algorithm; in addition, it has a fixed complexity and could be carried out by parallel processing.
List-Based Simulated Annealing Algorithm for Traveling Salesman Problem.
Zhan, Shi-hua; Lin, Juan; Zhang, Ze-jun; Zhong, Yi-wen
2016-01-01
Simulated annealing (SA) algorithm is a popular intelligent optimization algorithm which has been successfully applied in many fields. Parameters' setting is a key factor for its performance, but it is also a tedious work. To simplify parameters setting, we present a list-based simulated annealing (LBSA) algorithm to solve traveling salesman problem (TSP). LBSA algorithm uses a novel list-based cooling schedule to control the decrease of temperature. Specifically, a list of temperatures is created first, and then the maximum temperature in list is used by Metropolis acceptance criterion to decide whether to accept a candidate solution. The temperature list is adapted iteratively according to the topology of the solution space of the problem. The effectiveness and the parameter sensitivity of the list-based cooling schedule are illustrated through benchmark TSP problems. The LBSA algorithm, whose performance is robust on a wide range of parameter values, shows competitive performance compared with some other state-of-the-art algorithms.
Enterprise Human Resources Information Mining Based on Improved Apriori Algorithm
Lei He
2013-05-01
Full Text Available With the unceasing development of information and technology in today’s modern society, enterprises’ demand of human resources information mining is getting bigger and bigger. Based on the enterprise human resources information mining situation, this paper puts forward a kind of improved Apriori algorithm based model on the enterprise human resources information mining, this model introduced data mining technology and traditional Apriori algorithm, and improved on its basis, divided the association rules mining task of the original algorithm into two subtasks of producing frequent item sets and producing rule, using SQL technology to directly generating frequent item sets, and using the method of establishing chart to extract the information which are interested to customers. The experimental results show that the improved Apriori algorithm based model on the enterprise human resources information mining is better in efficiency than the original algorithm, and the practical application test results show that the improved algorithm is practical and effective.
Clonal Selection Algorithm Based Iterative Learning Control with Random Disturbance
Yuanyuan Ju
2013-01-01
Full Text Available Clonal selection algorithm is improved and proposed as a method to solve optimization problems in iterative learning control. And a clonal selection algorithm based optimal iterative learning control algorithm with random disturbance is proposed. In the algorithm, at the same time, the size of the search space is decreased and the convergence speed of the algorithm is increased. In addition a model modifying device is used in the algorithm to cope with the uncertainty in the plant model. In addition a model is used in the algorithm cope with the uncertainty in the plant model. Simulations show that the convergence speed is satisfactory regardless of whether or not the plant model is precise nonlinear plants. The simulation test verify the controlled system with random disturbance can reached to stability by using improved iterative learning control law but not the traditional control law.
An optimal scheduling algorithm based on task duplication
Ruan Youlin; Liu Gan; Zhu Guangxi; Lu Xiaofeng
2005-01-01
When the communication time is relatively shorter than the computation time for every task, the task duplication based scheduling (TDS) algorithm proposed by Darbha and Agrawal generates an optimal schedule. Park and Choe also proposed an extended TDS algorithm whose optimality condition is less restricted than that of TDS algorithm, but the condition is very complex and is difficult to satisfy when the number of tasks is large. An efficient algorithm is proposed whose optimality condition is less restricted and simpler than both of the algorithms, and the schedule length is also shorter than both of the algorithms. The time complexity of the proposed algorithm is O ( v2 ), where v represents the number of tasks.
OPTIMIZATION BASED ON LMPROVED REAL—CODED GENETIC ALGORITHM
ShiYu; YuShenglin
2002-01-01
An improved real-coded genetic algorithm is pro-posed for global optimization of functionsl.The new algo-rithm is based om the judgement of the searching perfor-mance of basic real-coded genetic algorithm.The opera-tions of basic real-coded genetic algorithm are briefly dis-cussed and selected.A kind of chaos sequence is described in detail and added in the new algorithm ad a disturbance factor.The strategy of field partition is also used to im-prove the strcture of the new algorithm.Numerical ex-periment shows that the mew genetic algorithm can find the global optimum of complex funtions with satistaiting precision.
Robust adaptive beamforming algorithm based on Bayesian approach
Xin SONG; Jinkuan WANG; Yinghua HAN; Han WANG
2008-01-01
The performance of adaptive array beamform-ing algorithms substantially degrades in practice because of a slight mismatch between actual and presumed array res-ponses to the desired signal. A novel robust adaptive beam-forming algorithm based on Bayesian approach is therefore proposed. The algorithm responds to the current envi-ronment by estimating the direction of arrival (DOA) of the actual signal from observations. Computational com-plexity of the proposed algorithm can thus be reduced com-pared with other algorithms since the recursive method is used to obtain inverse matrix. In addition, it has strong robustness to the uncertainty of actual signal DOA and makes the mean output array signal-to-interference-plus-noise ratio (SINR) consistently approach the optimum. Simulation results show that the proposed algorithm is bet-ter in performance than conventional adaptive beamform-ing algorithms.
New Gradient-Based Variable Step Size LMS Algorithms
Yanling Hao
2008-03-01
Full Text Available Two new gradient-based variable step size least-mean-square (VSSLMS algorithms are proposed on the basis of a concise assessment of the weaknesses of previous VSSLMS algorithms in high-measurement noise environments. The first algorithm is designed for applications where the measurement noise signal is statistically stationary and the second for statistically nonstationary noise. Steady-state performance analyses are provided for both algorithms and verified by simulations. The proposed algorithms are also confirmed by simulations to obtain both a fast convergence rate and a small steady-state excess mean square error (EMSE, and to outperform existing VSSLMS algorithms. To facilitate practical application, parameter choice guidelines are provided for the new algorithms.
Generating Decision Trees Method Based on Improved ID3 Algorithm
Yang Ming; Guo Shuxu1; Wang Jun3
2011-01-01
The ID3 algorithm is a classical learning algorithm of decision tree in data mining.The algorithm trends to choosing the attribute with more values,affect the efficiency of classification and prediction for building a decision tree.This article proposes a new approach based on an improved ID3 algorithm.The new algorithm introduces the importance factor λ when calculating the information entropy.It can strengthen the label of important attributes of a tree and reduce the label of non-important attributes.The algorithm overcomes the flaw of the traditional ID3 algorithm which tends to choose the attributes with more values,and also improves the efficiency and flexibility in the process of generating decision trees.
Fuzzy Rules for Ant Based Clustering Algorithm
Amira Hamdi
2016-01-01
Full Text Available This paper provides a new intelligent technique for semisupervised data clustering problem that combines the Ant System (AS algorithm with the fuzzy c-means (FCM clustering algorithm. Our proposed approach, called F-ASClass algorithm, is a distributed algorithm inspired by foraging behavior observed in ant colonyT. The ability of ants to find the shortest path forms the basis of our proposed approach. In the first step, several colonies of cooperating entities, called artificial ants, are used to find shortest paths in a complete graph that we called graph-data. The number of colonies used in F-ASClass is equal to the number of clusters in dataset. Hence, the partition matrix of dataset founded by artificial ants is given in the second step, to the fuzzy c-means technique in order to assign unclassified objects generated in the first step. The proposed approach is tested on artificial and real datasets, and its performance is compared with those of K-means, K-medoid, and FCM algorithms. Experimental section shows that F-ASClass performs better according to the error rate classification, accuracy, and separation index.
A POCS-Based Algorithm for Blocking Artifacts Reduction
ZHAO Yi-hong; CHENG Guo-hua; YU Song-yu
2006-01-01
An algorithm for blocking artifacts reduction in DCT domain for block-based image coding was developed. The algorithm is based on the projection onto convex set (POCS) theory. Due to the fact that the DCT characteristics of shifted blocks are different caused by the blocking artifacts, a novel smoothness constraint set and the corresponding projection operator were proposed to reduce the blocking artifacts by discarding the undesired high frequency coefficients in the shifted DCT blocks. The experimental results show that the proposed algorithm outperforms the conventional algorithms in terms of objective quality, subjective quality, and convergence property.
A Parallel Encryption Algorithm Based on Piecewise Linear Chaotic Map
Xizhong Wang
2013-01-01
Full Text Available We introduce a parallel chaos-based encryption algorithm for taking advantage of multicore processors. The chaotic cryptosystem is generated by the piecewise linear chaotic map (PWLCM. The parallel algorithm is designed with a master/slave communication model with the Message Passing Interface (MPI. The algorithm is suitable not only for multicore processors but also for the single-processor architecture. The experimental results show that the chaos-based cryptosystem possesses good statistical properties. The parallel algorithm provides much better performance than the serial ones and would be useful to apply in encryption/decryption file with large size or multimedia.
Heuristic Reduction Algorithm Based on Pairwise Positive Region
QI Li; LIU Yu-shu
2007-01-01
To guarantee the optimal reduct set, a heuristic reduction algorithm is proposed, which considers the distinguishing information between the members of each pair decision classes. Firstly the pairwise positive region is defined, based on which the pairwise significance measure is calculated between the members of each pair classes. Finally the weighted pairwise significance of attribute is used as the attribute reduction criterion, which indicates the necessity of attributes very well. By introducing the noise tolerance factor, the new algorithm can tolerate noise to some extent. Experimental results show the advantages of our novel heuristic reduction algorithm over the traditional attribute dependency based algorithm.
Survey of gene splicing algorithms based on reads.
Si, Xiuhua; Wang, Qian; Zhang, Lei; Wu, Ruo; Ma, Jiquan
2017-09-05
Gene splicing is the process of assembling a large number of unordered short sequence fragments to the original genome sequence as accurately as possible. Several popular splicing algorithms based on reads are reviewed in this article, including reference genome algorithms and de novo splicing algorithms (Greedy-extension, Overlap-Layout-Consensus graph, De Bruijn graph). We also discuss a new splicing method based on the MapReduce strategy and Hadoop. By comparing these algorithms, some conclusions are drawn and some suggestions on gene splicing research are made.
PHC: A Fast Partition and Hierarchy-Based Clustering Algorithm
ZHOU HaoFeng(周皓峰); YUAN QingQing(袁晴晴); CHENG ZunPing(程尊平); SHI BaiLe(施伯乐)
2003-01-01
Cluster analysis is a process to classify data in a specified data set. In this field,much attention is paid to high-efficiency clustering algorithms. In this paper, the features in thecurrent partition-based and hierarchy-based algorithms are reviewed, and a new hierarchy-basedalgorithm PHC is proposed by combining advantages of both algorithms, which uses the cohesionand the closeness to amalgamate the clusters. Compared with similar algorithms, the performanceof PHC is improved, and the quality of clustering is guaranteed. And both the features were provedby the theoretic and experimental analyses in the paper.
Brain MR image segmentation improved algorithm based on probability
Liao, Hengxu; Liu, Gang; Guo, Xiantang
2017-08-01
Local weight voting algorithm is a kind of current mainstream segmentation algorithm. It takes full account of the influences of the likelihood of image likelihood and the prior probabilities of labels on the segmentation results. But this method still can be improved since the essence of this method is to get the label with the maximum probability. If the probability of a label is 70%, it may be acceptable in mathematics. But in the actual segmentation, it may be wrong. So we use the matrix completion algorithm as a supplement. When the probability of the former is larger, the result of the former algorithm is adopted. When the probability of the later is larger, the result of the later algorithm is adopted. This is equivalent to adding an automatic algorithm selection switch that can theoretically ensure that the accuracy of the algorithm we propose is superior to the local weight voting algorithm. At the same time, we propose an improved matrix completion algorithm based on enumeration method. In addition, this paper also uses a multi-parameter registration model to reduce the influence that the registration made on the segmentation. The experimental results show that the accuracy of the algorithm is better than the common segmentation algorithm.
Adaptive image contrast enhancement algorithm for point-based rendering
Xu, Shaoping; Liu, Xiaoping P.
2015-03-01
Surgical simulation is a major application in computer graphics and virtual reality, and most of the existing work indicates that interactive real-time cutting simulation of soft tissue is a fundamental but challenging research problem in virtual surgery simulation systems. More specifically, it is difficult to achieve a fast enough graphic update rate (at least 30 Hz) on commodity PC hardware by utilizing traditional triangle-based rendering algorithms. In recent years, point-based rendering (PBR) has been shown to offer the potential to outperform the traditional triangle-based rendering in speed when it is applied to highly complex soft tissue cutting models. Nevertheless, the PBR algorithms are still limited in visual quality due to inherent contrast distortion. We propose an adaptive image contrast enhancement algorithm as a postprocessing module for PBR, providing high visual rendering quality as well as acceptable rendering efficiency. Our approach is based on a perceptible image quality technique with automatic parameter selection, resulting in a visual quality comparable to existing conventional PBR algorithms. Experimental results show that our adaptive image contrast enhancement algorithm produces encouraging results both visually and numerically compared to representative algorithms, and experiments conducted on the latest hardware demonstrate that the proposed PBR framework with the postprocessing module is superior to the conventional PBR algorithm and that the proposed contrast enhancement algorithm can be utilized in (or compatible with) various variants of the conventional PBR algorithm.
Research on Algorithms for Mining Distance-Based Outliers
WANGLizhen; ZOULikun
2005-01-01
The outlier detection is an important and valuable research in KDD (Knowledge discover in database). The identification of outliers can lead to the discovery of truly unexpected knowledge in areas such as electronic commerce, credit card fraud, and even weather forecast. In existing methods that we have seen for finding outliers, the notion of DB-(Distance-based) outliers is not restricted computationally to small values of the number of dimensions k and goes beyond the data space. Here, we study algorithms for mining DB-outliers. We focus on developing algorithms unlimited by k. First, we present a Partition-based algorithm (the PBA). The key idea is to gain efficiency by divide-and-conquer. Second, we present an optimized algorithm called Object-class-based algorithm (the OCBA). The computing of this algorithm has nothing to do with k and the efficiency of this algorithm is as good as the cell-based algorithm. We provide experimental results showing that the two new algorithms have better execution efficiency.
Grover quantum searching algorithm based on weighted targets
Li Panchi; Li Shiyong
2008-01-01
The current Grover quantum searching algorithm cannot identify the difference in importance of the search targets when it is applied to an unsorted quantum database, and the probability for each search target is equal. To solve this problem, a Grover searching algorithm based on weighted targets is proposed. First, each target is endowed a weight coefficient according to its importance. Applying these different weight coefficients, the targets are represented as quantum superposition states. Second, the novel Grover searching algorithm based on the quantum superposition of the weighted targets is constructed. Using this algorithm, the probability of getting each target can be approximated to the corresponding weight coefficient, which shows the flexibility of this algorithm.Finally, the validity of the algorithm is proved by a simple searching example.
Human resource recommendation algorithm based on ensemble learning and Spark
Cong, Zihan; Zhang, Xingming; Wang, Haoxiang; Xu, Hongjie
2017-08-01
Aiming at the problem of “information overload” in the human resources industry, this paper proposes a human resource recommendation algorithm based on Ensemble Learning. The algorithm considers the characteristics and behaviours of both job seeker and job features in the real business circumstance. Firstly, the algorithm uses two ensemble learning methods-Bagging and Boosting. The outputs from both learning methods are then merged to form user interest model. Based on user interest model, job recommendation can be extracted for users. The algorithm is implemented as a parallelized recommendation system on Spark. A set of experiments have been done and analysed. The proposed algorithm achieves significant improvement in accuracy, recall rate and coverage, compared with recommendation algorithms such as UserCF and ItemCF.
Lazy learner text categorization algorithm based on embedded feature selection
Yan Peng; Zheng Xuefeng; Zhu Jianyong; Xiao Yunhong
2009-01-01
To avoid the curse of dimensionality, text categorization (TC) algorithms based on machine learning (ML) have to use an feature selection (FS) method to reduce the dimensionality of feature space. Although having been widely used, FS process will generally cause information losing and then have much side-effect on the whole performance of TC algorithms. On the basis of the sparsity characteristic of text vectors, a new TC algorithm based on lazy feature selection (LFS) is presented. As a new type of embedded feature selection approach, the LFS method can greatly reduce the dimension of features without any information losing, which can improve both efficiency and performance of algorithms greatly. The experiments show the new algorithm can simultaneously achieve much higher both performance and efficiency than some of other classical TC algorithms.
无
2006-01-01
A novel method of global optimal path planning for mobile robot was proposed based on the improved Dijkstra algorithm and ant system algorithm. This method includes three steps: the first step is adopting the MAKLINK graph theory to establish the free space model of the mobile robot, the second step is adopting the improved Dijkstra algorithm to find out a sub-optimal collision-free path, and the third step is using the ant system algorithm to adjust and optimize the location of the sub-optimal path so as to generate the global optimal path for the mobile robot. The computer simulation experiment was carried out and the results show that this method is correct and effective. The comparison of the results confirms that the proposed method is better than the hybrid genetic algorithm in the global optimal path planning.
A real time vehicles detection algorithm for vision based sensors
Płaczek, Bartłomiej
2011-01-01
A vehicle detection plays an important role in the traffic control at signalised intersections. This paper introduces a vision-based algorithm for vehicles presence recognition in detection zones. The algorithm uses linguistic variables to evaluate local attributes of an input image. The image attributes are categorised as vehicle, background or unknown features. Experimental results on complex traffic scenes show that the proposed algorithm is effective for a real-time vehicles detection.
Algorithm Research of Individualized Travelling Route Recommendation Based on Similarity
Xue Shan; Liu Song
2015-01-01
Although commercial recommendation system has made certain achievement in travelling route development, the recommendation system is facing a series of challenges because of people’s increasing interest in travelling. It is obvious that the core content of the recommendation system is recommendation algorithm. The advantages of recommendation algorithm can bring great effect to the recommendation system. Based on this, this paper applies traditional collaborative filtering algorithm for analy...
A New RWA Algorithm Based on Multi-Objective
无
2003-01-01
In this article, we studied the associated research problems and challenges on routing and wavelength assignment (RWA) in WDM (wavelength division multiplexing) networks. Various RWA approaches are examined and compared. We proposed a new RWA algorithm based on multi-objective. In this new algorithm, we consider multiple network optimizing objectives to setup a lightpath with maximize profit and shortest path under the limited resources. By comparing and analyzing, the proposed algorithm is much better ...
Variable Neighborhood Search Based Algorithm for University Course Timetabling Problem
Kralev, Velin; Kraleva, Radoslava
2016-01-01
In this paper a variable neighborhood search approach as a method for solving combinatoric optimization problems is presented. A variable neighborhood search based algorithm for solving the problem concerning the university course timetable design has been developed. This algorithm is used to solve the real problem regarding the university course timetable design. It is compared with other algorithms that are tested on the same sets of input data. The object and the methodology of study are p...
TOA estimation algorithm based on multi-search
无
2005-01-01
A new time of arrival (TOA) estimation algorithm is proposed. The algorithm computes the optimal sub-correlation length based on the SNR theory. So the robust of TOA acquirement is guaranteed very well. Then, according to the actual transmission environment and network system, the multi-search method is given. From the simulation result,the algorithm shows a very high application value in the realization of wireless location system (WLS).
Variable Neighborhood Search Based Algorithm for University Course Timetabling Problem
Kralev, Velin; Kraleva, Radoslava
2016-01-01
In this paper a variable neighborhood search approach as a method for solving combinatoric optimization problems is presented. A variable neighborhood search based algorithm for solving the problem concerning the university course timetable design has been developed. This algorithm is used to solve the real problem regarding the university course timetable design. It is compared with other algorithms that are tested on the same sets of input data. The object and the methodology of study are p...
Hindi Parser-based on CKY algorithm
Nitin Hambir; Ambrish Srivastav
2012-01-01
Hindi parser is a tool which takes Hindi sentence and verifies whether or not given Hindi sentence is correct according to Hindi language grammar. Parsing is important for Natural Language Processing tools. Hindi parser uses the CKY (Coke- Kasami-Younger) parsing algorithm for Parsing of Hindi language. It parses whole sentence and generate a matrix
Intelligent Hybrid Cluster Based Classification Algorithm for Social Network Analysis
S. Muthurajkumar
2014-05-01
Full Text Available In this paper, we propose an hybrid clustering based classification algorithm based on mean approach to effectively classify to mine the ordered sequences (paths from weblog data in order to perform social network analysis. In the system proposed in this work for social pattern analysis, the sequences of human activities are typically analyzed by switching behaviors, which are likely to produce overlapping clusters. In this proposed system, a robust Modified Boosting algorithm is proposed to hybrid clustering based classification for clustering the data. This work is useful to provide connection between the aggregated features from the network data and traditional indices used in social network analysis. Experimental results show that the proposed algorithm improves the decision results from data clustering when combined with the proposed classification algorithm and hence it is proved that of provides better classification accuracy when tested with Weblog dataset. In addition, this algorithm improves the predictive performance especially for multiclass datasets which can increases the accuracy.
A Vehicle Detection Algorithm Based on Deep Belief Network
Hai Wang
2014-01-01
Full Text Available Vision based vehicle detection is a critical technology that plays an important role in not only vehicle active safety but also road video surveillance application. Traditional shallow model based vehicle detection algorithm still cannot meet the requirement of accurate vehicle detection in these applications. In this work, a novel deep learning based vehicle detection algorithm with 2D deep belief network (2D-DBN is proposed. In the algorithm, the proposed 2D-DBN architecture uses second-order planes instead of first-order vector as input and uses bilinear projection for retaining discriminative information so as to determine the size of the deep architecture which enhances the success rate of vehicle detection. On-road experimental results demonstrate that the algorithm performs better than state-of-the-art vehicle detection algorithm in testing data sets.
A PRESSURE-BASED ALGORITHM FOR CAVITATING FLOW COMPUTATIONS
ZHANG Ling-xin; ZHAO Wei-guo; SHAO Xue-ming
2011-01-01
A pressure-based algorithm for the prediction of cavitating flows is presented. The algorithm employs a set of equations including the Navier-Stokes equations and a cavitation model explaining the phase change between liquid and vapor. A pressure-based method is used to construct the algorithm and the coupling between pressure and velocity is considered. The pressure correction equation is derived from a new continuity equation which employs a source term related to phase change rate instead of the material derivative of density Dp/Dt.Thispressure-based algorithm allows for the computation of steady or unsteady,2-Dor 3-D cavitating flows. Two 2-D cases, flows around a flat-nose cylinder and around a NACA0015 hydrofoil, are simulated respectively, and the periodic cavitation behaviors associated with the re-entrant jets are captured. This algorithm shows good capability of computating time-dependent cavitating flows.
New Iterated Decoding Algorithm Based on Differential Frequency Hopping System
LIANG Fu-lin; LUO Wei-xiong
2005-01-01
A new iterated decoding algorithm is proposed for differential frequency hopping (DFH) encoder concatenated with multi-frequency shift-key (MFSK) modulator. According to the character of the frequency hopping (FH) pattern trellis produced by DFH function, maximum a posteriori (MAP) probability theory is applied to realize the iterate decoding of it. Further, the initial conditions for the new iterate algorithm based on MAP algorithm are modified for better performance. Finally, the simulation result compared with that from traditional algorithms shows good anti-interference performance.
Topology control based on quantum genetic algorithm in sensor networks
SUN Lijuan; GUO Jian; LU Kai; WANG Ruchuan
2007-01-01
Nowadays,two trends appear in the application of sensor networks in which both multi-service and quality of service (QoS)are supported.In terms of the goal of low energy consumption and high connectivity,the control on topology is crucial.The algorithm of topology control based on quantum genetic algorithm in sensor networks is proposed.An advantage of the quantum genetic algorithm over the conventional genetic algorithm is demonstrated in simulation experiments.The goals of high connectivity and low consumption of energy are reached.
Surname Inherited Algorithm Research Based on Artificial Immune System
Jing Xie
2013-06-01
Full Text Available To keep the diversity of antibodies in artificial immune system evolution process, this paper puts forward a kind of increase simulation surname inheritance algorithm based on the clonal selection algorithm, and identification and forecast the Vibration Data about CA6140 horizontal lathe machining slender shaft workpiece prone . The results show that the algorithm has the characteristics of flexible application, strong adaptability, an effective approach to improve efficiency of the algorithm, a good performance of global searching and broad application prospect.
Agent-based Algorithm for Spatial Distribution of Objects
Collier, Nathan
2012-06-02
In this paper we present an agent-based algorithm for the spatial distribution of objects. The algorithm is a generalization of the bubble mesh algorithm, initially created for the point insertion stage of the meshing process of the finite element method. The bubble mesh algorithm treats objects in space as bubbles, which repel and attract each other. The dynamics of each bubble are approximated by solving a series of ordinary differential equations. We present numerical results for a meshing application as well as a graph visualization application.
Support vector classification algorithm based on variable parameter linear programming
Xiao Jianhua; Lin Jian
2007-01-01
To solve the problems of SVM in dealing with large sample size and asymmetric distributed samples, a support vector classification algorithm based on variable parameter linear programming is proposed.In the proposed algorithm, linear programming is employed to solve the optimization problem of classification to decrease the computation time and to reduce its complexity when compared with the original model.The adjusted punishment parameter greatly reduced the classification error resulting from asymmetric distributed samples and the detailed procedure of the proposed algorithm is given.An experiment is conducted to verify whether the proposed algorithm is suitable for asymmetric distributed samples.
A multicast dynamic wavelength assignment algorithm based on matching degree
WU Qi-wu; ZHOU Xian-wei; WANG Jian-ping; YIN Zhi-hong; ZHANG Long
2009-01-01
The wavelength assignment with multiple multicast requests in fixed routing WDM network is studied. A new multicast dynamic wavelength assignment algorithm is presented based on matching degree. First, the wavelength matching degree between available wavelengths and multicast routing trees is introduced into the algorithm. Then, the wavelength assign-ment is translated into the maximum weight matching in bipartite graph, and this matching problem is solved by using an extended Kuhn-Munkres algorithm. The simulation results prove that the overall optimal wavelength assignment scheme is obtained in polynomial time. At the same time, the proposed algorithm can reduce the connecting blocking probability and improve the system resource utilization.
A new parallel algorithm for image matching based on entropy
董开坤; 胡铭曾
2001-01-01
Presents a new parallel image matching algorithm based on the concept of entropy feature vector and suitable to SIMD computer, which, in comparison with other algorithms, has the following advantages: ( 1 ) The spatial information of an image is appropriately introduced into the definition of image entropy. (2) A large number of multiplication operations are eliminated, thus the algorithm is sped up. (3) The shortcoming of having to do global calculation in the first instance is overcome, and concludes the algorithm has very good locality and is suitable for parallel processing.
The RSA Cryptoprocessor Hardware Implementation Based on Modified Montgomery Algorithm
CHEN Bo; WANG Xu; RONG Meng-tian
2005-01-01
RSA (Rivest-Shamir-Adleman)public-key cryptosystem is widely used in the information security area such as encryption and digital signature. Based on the modified Montgomery modular multiplication algorithm, a new architecture using CSA(carry save adder)was presented to implement modular multiplication. Compared with the popular modular multiplication algorithms using two CSA, the presented algorithm uses only one CSA, so it can improve the time efficiency of RSA cryptoprocessor and save about half of hardware resources for modular multiplication. With the increase of encryption data size n, the clock cycles for the encryption procedure reduce in T(n2) , compared with the modular multiplication algorithms using two CSA.
An Incremental Rule Acquisition Algorithm Based on Rough Set
YU Hong; YANG Da-chun
2005-01-01
Rough Set is a valid mathematical theory developed in recent years,which has the ability to deal with imprecise,uncertain,and vague information.This paper presents a new incremental rule acquisition algorithm based on rough set theory.First,the relation of the new instances with the original rule set is discussed.Then the change law of attribute reduction and value reduction are studied when a new instance is added.Follow,a new incremental learning algorithm for decision tables is presented within the framework of rough set.Finally,the new algorithm and the classical algorithm are analyzed and compared by theory and experiments.
CUDT: a CUDA based decision tree algorithm.
Lo, Win-Tsung; Chang, Yue-Shan; Sheu, Ruey-Kai; Chiu, Chun-Chieh; Yuan, Shyan-Ming
2014-01-01
Decision tree is one of the famous classification methods in data mining. Many researches have been proposed, which were focusing on improving the performance of decision tree. However, those algorithms are developed and run on traditional distributed systems. Obviously the latency could not be improved while processing huge data generated by ubiquitous sensing node in the era without new technology help. In order to improve data processing latency in huge data mining, in this paper, we design and implement a new parallelized decision tree algorithm on a CUDA (compute unified device architecture), which is a GPGPU solution provided by NVIDIA. In the proposed system, CPU is responsible for flow control while the GPU is responsible for computation. We have conducted many experiments to evaluate system performance of CUDT and made a comparison with traditional CPU version. The results show that CUDT is 5 ∼ 55 times faster than Weka-j48 and is 18 times speedup than SPRINT for large data set.
Novel Adaptive Beamforming Algorithm Based on Wavelet Packet Transform
Zhang Xiaofei; Xu Dazhuan
2005-01-01
An analysis of the received signal of array antennas shows that the received signal has multi-resolution characteristics, and hence the wavelet packet theory can be used to detect the signal. By emplying wavelet packet theory to adaptive beamforming, a wavelet packet transform-based adaptive beamforming algorithm (WP-ABF) is proposed . This WP-ABF algorithm uses wavelet packet transform as the preprocessing, and the wavelet packet transformed signal uses least mean square algorithm to implement the adaptive beamforming. White noise can be wiped off under wavelet packet transform according to the different characteristics of signal and white under the wavelet packet transform. Theoretical analysis and simulations demonstrate that the proposed WP-ABF algorithm converges faster than the conventional adaptive beamforming algorithm and the wavelet transform-based beamforming algorithm. Simulation results also reveal that the convergence of the algorithm relates closely to the wavelet base and series; that is, the algorithm convergence gets better with the increasing of series, and for the same series of wavelet base the convergence gets better with the increasing of regularity.
New MPPT algorithm based on hybrid dynamical theory
Elmetennani, Shahrazed
2014-11-01
This paper presents a new maximum power point tracking algorithm based on the hybrid dynamical theory. A multiceli converter has been considered as an adaptation stage for the photovoltaic chain. The proposed algorithm is a hybrid automata switching between eight different operating modes, which has been validated by simulation tests under different working conditions. © 2014 IEEE.
A danger-theory-based immune network optimization algorithm.
Zhang, Ruirui; Li, Tao; Xiao, Xin; Shi, Yuanquan
2013-01-01
Existing artificial immune optimization algorithms reflect a number of shortcomings, such as premature convergence and poor local search ability. This paper proposes a danger-theory-based immune network optimization algorithm, named dt-aiNet. The danger theory emphasizes that danger signals generated from changes of environments will guide different levels of immune responses, and the areas around danger signals are called danger zones. By defining the danger zone to calculate danger signals for each antibody, the algorithm adjusts antibodies' concentrations through its own danger signals and then triggers immune responses of self-regulation. So the population diversity can be maintained. Experimental results show that the algorithm has more advantages in the solution quality and diversity of the population. Compared with influential optimization algorithms, CLONALG, opt-aiNet, and dopt-aiNet, the algorithm has smaller error values and higher success rates and can find solutions to meet the accuracies within the specified function evaluation times.
Analog Circuit Design Optimization Based on Evolutionary Algorithms
Mansour Barari
2014-01-01
Full Text Available This paper investigates an evolutionary-based designing system for automated sizing of analog integrated circuits (ICs. Two evolutionary algorithms, genetic algorithm and PSO (Parswal particle swarm optimization algorithm, are proposed to design analog ICs with practical user-defined specifications. On the basis of the combination of HSPICE and MATLAB, the system links circuit performances, evaluated through specific electrical simulation, to the optimization system in the MATLAB environment, for the selected topology. The system has been tested by typical and hard-to-design cases, such as complex analog blocks with stringent design requirements. The results show that the design specifications are closely met. Comparisons with available methods like genetic algorithms show that the proposed algorithm offers important advantages in terms of optimization quality and robustness. Moreover, the algorithm is shown to be efficient.
A Practical Localization Algorithm Based on Wireless Sensor Networks
Huang, Tao; Xia, Feng; Jin, Cheng; Li, Liang
2010-01-01
Many localization algorithms and systems have been developed by means of wireless sensor networks for both indoor and outdoor environments. To achieve higher localization accuracy, extra hardware equipments are utilized by most of the existing localization algorithms, which increase the cost and greatly limit the range of location-based applications. In this paper we present a method which can effectively meet different localization accuracy requirements of most indoor and outdoor location services in realistic applications. Our algorithm is composed of two phases: partition phase, in which the target region is split into small grids and localization refinement phase in which a higher accuracy location can be generated by applying a trick algorithm. A realistic demo system using our algorithm has been developed to illustrate its feasibility and availability. The results show that our algorithm can improve the localization accuracy.
Teaching learning based optimization algorithm and its engineering applications
Rao, R Venkata
2016-01-01
Describing a new optimization algorithm, the “Teaching-Learning-Based Optimization (TLBO),” in a clear and lucid style, this book maximizes reader insights into how the TLBO algorithm can be used to solve continuous and discrete optimization problems involving single or multiple objectives. As the algorithm operates on the principle of teaching and learning, where teachers influence the quality of learners’ results, the elitist version of TLBO algorithm (ETLBO) is described along with applications of the TLBO algorithm in the fields of electrical engineering, mechanical design, thermal engineering, manufacturing engineering, civil engineering, structural engineering, computer engineering, electronics engineering, physics and biotechnology. The book offers a valuable resource for scientists, engineers and practitioners involved in the development and usage of advanced optimization algorithms.
Acceleration of Directional Medain Filter Based Deinterlacing Algorithm (DMFD
Addanki Purna Ramesh
2011-12-01
Full Text Available This paper presents a novel directional median filter based deinterlacing algorithm (DMFD. DMFD is a content adaptive spatial deinterlacing algorithm that finds the direction of the edge and applies the median filtering along the edge to interpolate the odd pixels from the 5 pixels from the upper and 5 pixels from the lower even lines of the field. The proposed algorithm gives a significance improvement of 3db for baboon standard test image that has high textured content compared to CADEM, DOI, and MELA and also gives improved average PSNR compared previous algorithms. The algorithm written and tested in C and ported onto Altera’s NIOS II embedded soft processor and configured in CYCLONE-II FPGA. The ISA of Nios-II processor has extended with two additional instructions for calculation of absolute difference and minimum of four numbers to accelerate the FPGA implementation of the algorithms by 3.2 times
Local Community Detection Algorithm Based on Minimal Cluster
Yong Zhou
2016-01-01
Full Text Available In order to discover the structure of local community more effectively, this paper puts forward a new local community detection algorithm based on minimal cluster. Most of the local community detection algorithms begin from one node. The agglomeration ability of a single node must be less than multiple nodes, so the beginning of the community extension of the algorithm in this paper is no longer from the initial node only but from a node cluster containing this initial node and nodes in the cluster are relatively densely connected with each other. The algorithm mainly includes two phases. First it detects the minimal cluster and then finds the local community extended from the minimal cluster. Experimental results show that the quality of the local community detected by our algorithm is much better than other algorithms no matter in real networks or in simulated networks.
Compressive sensing based algorithms for electronic defence
Mishra, Amit Kumar
2017-01-01
This book details some of the major developments in the implementation of compressive sensing in radio applications for electronic defense and warfare communication use. It provides a comprehensive background to the subject and at the same time describes some novel algorithms. It also investigates application value and performance-related parameters of compressive sensing in scenarios such as direction finding, spectrum monitoring, detection, and classification.
Image completion algorithm based on texture synthesis
Zhang Hongying; Peng Qicong; Wu Yadong
2007-01-01
A new algorithm is proposed for completing the missing parts caused by the removal of foreground or background elements from an image of natural scenery in a visually plausible way.The major contributions of the proposed algorithm are: (1) for most natural images, there is a strong orientation of texture or color distribution.So a method is introduced to compute the main direction of the texture and complete the image by limiting the search to one direction to carry out image completion quite fast; (2) there exists a synthesis ordering for image completion.The searching order of the patches is denned to ensure the regions with more known information and the structures should be completed before filling in other regions; (3) to improve the visual effect of texture synthesis, an adaptive scheme is presented to determine the size of the template window for capturing the features of various scales.A number of examples are given to demonstrate the effectiveness of the proposed algorithm.
A Novel Heuristic Algorithm Based on Clark and Wright Algorithm for Green Vehicle Routing Problem
Mehdi Alinaghian; Zahra Kaviani; Siyavash Khaledan
2015-01-01
A significant portion of Gross Domestic Production (GDP) in any country belongs to the transportation system. Transportation equipment, in the other hand, is supposed to be great consumer of oil products. Many attempts have been assigned to the vehicles to cut down Greenhouse Gas (GHG). In this paper a novel heuristic algorithm based on Clark and Wright Algorithm called Green Clark and Wright (GCW) for Vehicle Routing Problem regarding to fuel consumption is presented. The objective function ...
Collaborative Filtering Algorithms Based on Kendall Correlation in Recommender Systems
YAO Yu; ZHU Shanfeng; CHEN Xinmeng
2006-01-01
In this work, Kendall correlation based collaborative filtering algorithms for the recommender systems are proposed. The Kendall correlation method is used to measure the correlation amongst users by means of considering the relative order of the users' ratings. Kendall based algorithm is based upon a more general model and thus could be more widely applied in e-commerce. Another discovery of this work is that the consideration of only positive correlated neighbors in prediction, in both Pearson and Kendall algorithms, achieves higher accuracy than the consideration of all neighbors, with only a small loss of coverage.
Video segmentation using multiple features based on EM algorithm
张风超; 杨杰; 刘尔琦
2004-01-01
Object-based video segmentation is an important issue for many multimedia applications. A video segmentation method based on EM algorithm is proposed. We consider video segmentation as an unsupervised classification problem and apply EM algorithm to obtain the maximum-likelihood estimation of the Gaussian model parameters for model-based segmentation. We simultaneously combine multiple features (motion, color) within a maximum likelihood framework to obtain accurate segment results. We also use the temporal consistency among video frames to improve the speed of EM algorithm. Experimental results on typical MPEG-4 sequences and real scene sequences show that our method has an attractive accuracy and robustness.
Fast image matching algorithm based on projection characteristics
Zhou, Lijuan; Yue, Xiaobo; Zhou, Lijun
2011-06-01
Based on analyzing the traditional template matching algorithm, this paper identified the key factors restricting the speed of matching and put forward a brand new fast matching algorithm based on projection. Projecting the grayscale image, this algorithm converts the two-dimensional information of the image into one-dimensional one, and then matches and identifies through one-dimensional correlation, meanwhile, because of normalization has been done, when the image brightness or signal amplitude increasing in proportion, it could also perform correct matching. Experimental results show that the projection characteristics based image registration method proposed in this article could greatly improve the matching speed, which ensuring the matching accuracy as well.
Research of Collaborative Filtering Recommendation Algorithm based on Network Structure
Hui PENG
2013-10-01
Full Text Available This paper combines the classic collaborative filtering algorithm with personalized recommendation algorithm based on network structure. For the data sparsity and malicious behavior problems of traditional collaborative filtering algorithm, the paper introduces a new kind of social network-based collaborative filtering algorithm. In order to improve the accuracy of the personalized recommendation technology, we first define empty state in the state space of multi-dimensional semi-Markov processes and obtain extended multi-dimensional semi-Markov processes which are combined with social network analysis theory, and then we get social network information flow model. The model describes the flow of information between the members of the social network. So, we propose collaborative filtering algorithm based on social network information flow model. The algorithm uses social network information and combines user trust with user interest and find nearest neighbors of the target user and then forms a project recommended to improve the accuracy of recommended. Compared with the traditional collaborative filtering algorithm, the algorithm can effectively alleviate the sparsity and malicious behavior problem, and significantly improve the quality of the recommendation system recommended.
Area Variation Based Color Snake Algorithm for Moving Object Tracking
Shoum-ik ROYCHOUDHURY; Young-joon HAN
2010-01-01
A snake algorithm has been known that it has a strong point in extracting the exact contour of an object.But it is apt to be influenced by scattered edges around the control points.Since the shape of a moving object in 2D image changes a lot due ta its rotation and translation in the 3D space,the conventional algorithm that takes into account slowly moving objects cannot provide an appropriate solution.To utilize the advantages of the snake algrithm while minimizing the drawbacks,this paper proposes the area variation based color snake algorithm for moving object tracking.The proposed algorithm includes a new energy term which is used for preserving the shape of an object between two consecutive inages.The proposed one can also segment precisely interesting objects on complex image since it is based on color information.Experiment results show that the proposed algorithm is very effective in various environments.
CUDA Based Speed Optimization of the PCA Algorithm
Salih Görgünoğlu
2016-05-01
Full Text Available Principal Component Analysis (PCA is an algorithm involving heavy mathematical operations with matrices. The data extracted from the face images are usually very large and to process this data is time consuming. To reduce the execution time of these operations, parallel programming techniques are used. CUDA is a multipurpose parallel programming architecture supported by graphics cards. In this study we have implemented the PCA algorithm using both the classical programming approach and CUDA based implementation using different configurations. The algorithm is subdivided into its constituent calculation steps and evaluated for the positive effects of parallelization on each step. Therefore, the parts of the algorithm that cannot be improved by parallelization are identified. On the other hand, it is also shown that, with CUDA based approach dramatic improvements in the overall performance of the algorithm arepossible.
Novel algorithm for distributed replicas management based on dynamic programming
Wang Tao; Lu Xianliang; Hou Mengshu
2006-01-01
Replicas can improve the data reliability in distributed system. However, the traditional algorithms for replica management are based on the assumption that all replicas have the uniform reliability, which is inaccurate in some actual systems. To address such problem, a novel algorithm is proposed based on dynamic programming to manage the number and distribution of replicas in different nodes. By using Markov model, replicas management is organized as a multi-phase process, and the recursion equations are provided. In this algorithm, the heterogeneity of nodes, the expense for maintaining replicas and the engaged space have been considered. Under these restricted conditions, this algorithm realizes high data reliability in a distributed system. The results of case analysis prove the feasibility of the algorithm.
Heuristic based data scheduling algorithm for OFDMA wireless network
无
2008-01-01
A system model based on joint layer mechanism is formulated for optimal data scheduling over fixed point-to-point links in OFDMA ad-hoc wireless networks.A distributed scheduling algorithm (DSA) for system model optimization is proposed that combines the randomly chosen subcarrier according to the channel condition of local subcarriers with link power control to limit interference caused by the reuse of subcarrier among links.For the global fairness improvement of algorithms,a global power control scheduling algorithm (GPCSA) based on the proposed DSA is presented and dynamically allocates global power according to difference between average carrier-noise-ratio of selected local links and system link protection ratio.Simulation results demonstrate that the proposed algorithms achieve better efficiency and fairness compared with other existing algorithms.
Drilling Path Optimization Based on Particle Swarm Optimization Algorithm
ZHU Guangyu; ZHANG Weibo; DU Yuexiang
2006-01-01
This paper presents a new approach based on the particle swarm optimization (PSO) algorithm for solving the drilling path optimization problem belonging to discrete space. Because the standard PSO algorithm is not guaranteed to be global convergence or local convergence, based on the mathematical algorithm model, the algorithm is improved by adopting the method of generate the stop evolution particle over again to get the ability of convergence to the global optimization solution. And the operators are improved by establishing the duality transposition method and the handle manner for the elements of the operator, the improved operator can satisfy the need of integer coding in drilling path optimization. The experiment with small node numbers indicates that the improved algorithm has the characteristics of easy realize, fast convergence speed, and better global convergence characteristics, hence the new PSO can play a role in solving the problem of drilling path optimization in drilling holes.
Image Recovery Algorithm Based on Learned Dictionary
Xinghui Zhu
2014-01-01
Full Text Available We proposed a recovery scheme for image deblurring. The scheme is under the framework of sparse representation and it has three main contributions. Firstly, considering the sparse property of natural image, the nonlocal overcompleted dictionaries are learned for image patches in our scheme. And, then, we coded the patches in each nonlocal clustering with the corresponding learned dictionary to recover the whole latent image. In addition, for some practical applications, we also proposed a method to evaluate the blur kernel to make the algorithm usable in blind image recovery. The experimental results demonstrated that the proposed scheme is competitive with some current state-of-the-art methods.
A Novel Heuristic Algorithm Based on Clark and Wright Algorithm for Green Vehicle Routing Problem
Mehdi Alinaghian
2015-08-01
Full Text Available A significant portion of Gross Domestic Production (GDP in any country belongs to the transportation system. Transportation equipment, in the other hand, is supposed to be great consumer of oil products. Many attempts have been assigned to the vehicles to cut down Greenhouse Gas (GHG. In this paper a novel heuristic algorithm based on Clark and Wright Algorithm called Green Clark and Wright (GCW for Vehicle Routing Problem regarding to fuel consumption is presented. The objective function is fuel consumption, drivers, and the usage of vehicles. Being compared to exact methods solutions for small-sized problems and to Differential Evolution (DE algorithm solutions for large-scaled problems, the results show efficient performance of the proposed GCW algorithm.
Real-Coded Quantum-Inspired Genetic Algorithm-Based BP Neural Network Algorithm
Jianyong Liu
2015-01-01
Full Text Available The method that the real-coded quantum-inspired genetic algorithm (RQGA used to optimize the weights and threshold of BP neural network is proposed to overcome the defect that the gradient descent method makes the algorithm easily fall into local optimal value in the learning process. Quantum genetic algorithm (QGA is with good directional global optimization ability, but the conventional QGA is based on binary coding; the speed of calculation is reduced by the coding and decoding processes. So, RQGA is introduced to explore the search space, and the improved varied learning rate is adopted to train the BP neural network. Simulation test shows that the proposed algorithm is effective to rapidly converge to the solution conformed to constraint conditions.
CUDT: A CUDA Based Decision Tree Algorithm
Win-Tsung Lo
2014-01-01
Full Text Available Decision tree is one of the famous classification methods in data mining. Many researches have been proposed, which were focusing on improving the performance of decision tree. However, those algorithms are developed and run on traditional distributed systems. Obviously the latency could not be improved while processing huge data generated by ubiquitous sensing node in the era without new technology help. In order to improve data processing latency in huge data mining, in this paper, we design and implement a new parallelized decision tree algorithm on a CUDA (compute unified device architecture, which is a GPGPU solution provided by NVIDIA. In the proposed system, CPU is responsible for flow control while the GPU is responsible for computation. We have conducted many experiments to evaluate system performance of CUDT and made a comparison with traditional CPU version. The results show that CUDT is 5∼55 times faster than Weka-j48 and is 18 times speedup than SPRINT for large data set.
A NOVEL THRESHOLD BASED EDGE DETECTION ALGORITHM
Y. RAMADEVI,
2011-06-01
Full Text Available Image segmentation is the process of partitioning/subdividing a digital image into multiple meaningful regions or sets of pixels regions with respect to a particular application. Edge detection is one of the frequently used techniques in digital image processing. The level to which the subdivision is carried depends on theproblem being viewed. Edges characterize boundaries and are therefore a problem of fundamental importance in image processing. There are many ways to perform edge detection. In this paper different Edge detection methods such as Sobel, Prewitt, Robert, Canny, Laplacian of Gaussian (LOG are used for segmenting the image. Expectation-Maximization (EM algorithm, OSTU and Genetic algorithms are also used. A new edge detection technique is proposed which detects the sharp and accurate edges that are not possible with the existing techniques. The proposed method with different threshold values for given input image is shown that ranges between 0 and 1 and it are observed that when the threshold value is 0.68 the sharp edges are recognised properly.
Efficient Satellite Scheduling Based on Improved Vector Evaluated Genetic Algorithm
Tengyue Mao
2012-03-01
Full Text Available Satellite scheduling is a typical multi-peak, many-valley, nonlinear multi-objective optimization problem. How to effectively implement the satellite scheduling is a crucial research in space areas.This paper mainly discusses the performance of VEGA (Vector Evaluated Genetic Algorithm based on the study of basic principles of VEGA algorithm, algorithm realization and test function, and then improves VEGA algorithm through introducing vector coding, new crossover and mutation operators, new methods to assign fitness and hold good individuals. As a result, the diversity and convergence of improved VEGA algorithm of improved VEGA algorithm have been significantly enhanced and will be applied to Earth-Mars orbit optimization. At the same time, this paper analyzes the results of the improved VEGA, whose results of performance analysis and evaluation show that although VEGA has a profound impact upon multi-objective evolutionary research, multi-objective evolutionary algorithm on the basis of Pareto seems to be a more effective method to get the non-dominated solutions from the perspective of diversity and convergence of experimental result. Finally, based on Visual C + + integrated development environment, we have implemented improved vector evaluation algorithm in the satellite scheduling.
ROUTING AND WAVELENGTH ASSIGNMENT ALGORITHMS BASED ON EQUIVALENT NETWORKS
Qi Xiaogang; Liu Lifang; Lin Sanyang
2006-01-01
In this paper, a Wavelength Division Multiplexing (WDM) network model based on the equivalent networks is described, and wavelength-dependent equivalent arc, equivalent networks, equivalent multicast tree and some other terms are presented. Based on this model and relevant Routing and Wavelength Assignment (RWA) strategy, a unicast RWA algorithm and a multicast RWA algorithm are presented. The wavelength-dependent equivalent arc expresses the schedule of local RWA and the equivalent network expresses the whole topology of WDM optical networks, so the two algorithms are of the flexibility in RWA and the optimization of the whole problem. The theoretic analysis and simulation results show the two algorithms are of the stronger capability and the lower complexity than the other existing algorithms for RWA problem, and the complexity of the two algorithms are only related to the scale of the equivalent networks. Finally, we prove the two algorithms' feasibility and the one-by-one corresponding relation between the equivalent multicast tree and original multicast tree, and point out the superiorities and drawbacks of the two algorithms respectively.
A novel bit-quad-based Euler number computing algorithm.
Yao, Bin; He, Lifeng; Kang, Shiying; Chao, Yuyan; Zhao, Xiao
2015-01-01
The Euler number of a binary image is an important topological property in computer vision and pattern recognition. This paper proposes a novel bit-quad-based Euler number computing algorithm. Based on graph theory and analysis on bit-quad patterns, our algorithm only needs to count two bit-quad patterns. Moreover, by use of the information obtained during processing the previous bit-quad, the average number of pixels to be checked for processing a bit-quad is only 1.75. Experimental results demonstrated that our method outperforms significantly conventional Euler number computing algorithms.
[Heart rate measurement algorithm based on artificial intelligence].
Chengxian, Cai; Wei, Wang
2010-01-01
Based on the heart rate measurement method using time-lapse image of human cheek, this paper proposes a novel measurement algorithm based on Artificial Intelligence. The algorithm combining with fuzzy logic theory acquires the heart beat point by using the defined fuzzy membership function of each sampled point. As a result, it calculates the heart rate by counting the heart beat points in a certain time period. Experiment shows said algorithm satisfies in operability, accuracy and robustness, which leads to constant practical value.
THE PARALLEL RECURSIVE AP ADAPTIVE ALGORITHM BASED ON VOLTERRA SERIES
无
2005-01-01
Aiming at the nonlinear system identification problem, a parallel recursive affine projection (AP) adaptive algorithm for the nonlinear system based on Volterra series is presented in this paper. The algorithm identifies in parallel the Volterra kernel of each order, recursively estimate the inverse of the autocorrelation matrix for the Volterra input of each order, and remarkably improve the convergence speed of the identification process compared with the NLMS and conventional AP adaptive algorithm based on Volterra series. Simulation results indicate that the proposed method in this paper is efficient.
Mercer Kernel Based Fuzzy Clustering Self-Adaptive Algorithm
李侃; 刘玉树
2004-01-01
A novel mercer kernel based fuzzy clustering self-adaptive algorithm is presented. The mercer kernel method is introduced to the fuzzy c-means clustering. It may map implicitly the input data into the high-dimensional feature space through the nonlinear transformation. Among other fuzzy c-means and its variants, the number of clusters is first determined. A self-adaptive algorithm is proposed. The number of clusters, which is not given in advance, can be gotten automatically by a validity measure function. Finally, experiments are given to show better performance with the method of kernel based fuzzy c-means self-adaptive algorithm.
A Novel Approach to Fast Image Filtering Algorithm of Infrared Images based on Intro Sort Algorithm
Gupta, Kapil Kumar; Niranjan, Jitendra Kumar
2012-01-01
In this study we investigate the fast image filtering algorithm based on Intro sort algorithm and fast noise reduction of infrared images. Main feature of the proposed approach is that no prior knowledge of noise required. It is developed based on Stefan- Boltzmann law and the Fourier law. We also investigate the fast noise reduction approach that has advantage of less computation load. In addition, it can retain edges, details, text information even if the size of the window increases. Intro sort algorithm begins with Quick sort and switches to heap sort when the recursion depth exceeds a level based on the number of elements being sorted. This approach has the advantage of fast noise reduction by reducing the comparison time. It also significantly speed up the noise reduction process and can apply to real-time image processing. This approach will extend the Infrared images applications for medicine and video conferencing.
Analysis of a wavelet-based robust hash algorithm
Meixner, Albert; Uhl, Andreas
2004-06-01
This paper paper is a quantitative evaluation of a wavelet-based, robust authentication hashing algorithm. Based on the results of a series of robustness and tampering sensitivity tests, we describepossible shortcomings and propose variousmodifications to the algorithm to improve its performance. The second part of the paper describes and attack against the scheme. It allows an attacker to modify a tampered image, such that it's hash value closely matches the hash value of the original.
An Event Grouping Based Algorithm for University Course Timetabling Problem
Kralev, Velin; Kraleva, Radoslava; Yurukov, Borislav
2016-01-01
This paper presents the study of an event grouping based algorithm for a university course timetabling problem. Several publications which discuss the problem and some approaches for its solution are analyzed. The grouping of events in groups with an equal number of events in each group is not applicable to all input data sets. For this reason, a universal approach to all possible groupings of events in commensurate in size groups is proposed here. Also, an implementation of an algorithm base...
An Event Grouping Based Algorithm for University Course Timetabling Problem
Kralev, Velin; Kraleva, Radoslava; Yurukov, Borislav
2016-01-01
This paper presents the study of an event grouping based algorithm for a university course timetabling problem. Several publications which discuss the problem and some approaches for its solution are analyzed. The grouping of events in groups with an equal number of events in each group is not applicable to all input data sets. For this reason, a universal approach to all possible groupings of events in commensurate in size groups is proposed here. Also, an implementation of an algorithm base...
Genetic Algorithm Based Microscale Vehicle Emissions Modelling
Sicong Zhu
2015-01-01
Full Text Available There is a need to match emission estimations accuracy with the outputs of transport models. The overall error rate in long-term traffic forecasts resulting from strategic transport models is likely to be significant. Microsimulation models, whilst high-resolution in nature, may have similar measurement errors if they use the outputs of strategic models to obtain traffic demand predictions. At the microlevel, this paper discusses the limitations of existing emissions estimation approaches. Emission models for predicting emission pollutants other than CO2 are proposed. A genetic algorithm approach is adopted to select the predicting variables for the black box model. The approach is capable of solving combinatorial optimization problems. Overall, the emission prediction results reveal that the proposed new models outperform conventional equations in terms of accuracy and robustness.
Warehouse Optimization Model Based on Genetic Algorithm
Guofeng Qin
2013-01-01
Full Text Available This paper takes Bao Steel logistics automated warehouse system as an example. The premise is to maintain the focus of the shelf below half of the height of the shelf. As a result, the cost time of getting or putting goods on the shelf is reduced, and the distance of the same kind of goods is also reduced. Construct a multiobjective optimization model, using genetic algorithm to optimize problem. At last, we get a local optimal solution. Before optimization, the average cost time of getting or putting goods is 4.52996 s, and the average distance of the same kinds of goods is 2.35318 m. After optimization, the average cost time is 4.28859 s, and the average distance is 1.97366 m. After analysis, we can draw the conclusion that this model can improve the efficiency of cargo storage.
Evolutionary Algorithm Based on Immune Strategy
WANG Lei; JIAO Licheng
2001-01-01
A novel evolutionary algorithm,evolution-immunity strategies(EIS), is proposed with reference to the immune theory in biology, which constructs an immune operator accomplished by two steps, a vaccination and an immune selection. The aim of introducing the immune concepts and methods into ES is for finding the ways and means obtaining the optimal solution of difficult problems with locally characteristic information. The detail processes of realizing EIS are presented which contain 6 steps. EIS is analyzed with Markovian theory and it is approved to be convergent with probability 1. In EIS, an immune operator is an aggregation of specific operations and procedures, and methods of selecting vaccines and constructing an immune operator are given in this paper. It is shown with an example of the 442-city TSP that the EIS can restrain the degenerate phenomenon during the evolutionary process by simulated calculating result, improve the searching capability and efficiency, and therefore, increase the convergent speed greatly.
RELAY ALGORITHM BASED ON NETWORK CODING IN WIRELESS LOCAL NETWORK
Wang Qi; Wang Qingshan; Wang Dongxue
2013-01-01
The network coding is a new technology in the field of information in 21st century.It could enhance the network throughput and save the energy consumption,and is mainly based on the single transmission rate.However,with the development of wireless network and equipment,wireless local network MAC protocols have already supported the multi-rate transmission.This paper investigates the optimal relay selection problem based on network coding.Firstly,the problem is formulated as an optimization problem.Moreover,a relay algorithm based on network coding is proposed and the transmission time gain of our algorithm over the traditional relay algorithm is analyzed.Lastly,we compare total transmission time and the energy consumption of our proposed algorithm,Network Coding with Relay Assistance (NCRA),Transmission Request (TR),and the Direct Transmission (DT) without relay algorithm by adopting IEEE 802.11b.The simulation results demonstrate that our algorithm that improves the coding opportunity by the cooperation of the relay nodes leads to the transmission time decrease of up to 17％ over the traditional relay algorithms.
Haplotyping a single triploid individual based on genetic algorithm.
Wu, Jingli; Chen, Xixi; Li, Xianchen
2014-01-01
The minimum error correction model is an important combinatorial model for haplotyping a single individual. In this article, triploid individual haplotype reconstruction problem is studied by using the model. A genetic algorithm based method GTIHR is presented for reconstructing the triploid individual haplotype. A novel coding method and an effectual hill-climbing operator are introduced for the GTIHR algorithm. This relatively short chromosome code can lead to a smaller solution space, which plays a positive role in speeding up the convergence process. The hill-climbing operator ensures algorithm GTIHR converge at a good solution quickly, and prevents premature convergence simultaneously. The experimental results prove that algorithm GTIHR can be implemented efficiently, and can get higher reconstruction rate than previous algorithms.
Distribution network planning algorithm based on Hopfield neural network
GAO Wei-xin; LUO Xian-jue
2005-01-01
This paper presents a new algorithm based on Hopfield neural network to find the optimal solution for an electric distribution network. This algorithm transforms the distribution power network-planning problem into a directed graph-planning problem. The Hopfield neural network is designed to decide the in-degree of each node and is in combined application with an energy function. The new algorithm doesn't need to code city streets and normalize data, so the program is easier to be realized. A case study applying the method to a district of 29 street proved that an optimal solution for the planning of such a power system could be obtained by only 26 iterations. The energy function and algorithm developed in this work have the following advantages over many existing algorithms for electric distribution network planning: fast convergence and unnecessary to code all possible lines.
Switching Equalization Algorithm Based on a New SNR Estimation Method
无
2007-01-01
It is well-known that turbo equalization with the max-log-map (MLM) rather than the log-map (LM) algorithm is insensitive to signal to noise ratio (SNR) mismatch. As our first contribution, an improved MLM algorithm called scaled max-log-map (SMLM) algorithm is presented. Simulation results show that the SMLM scheme can dramatically outperform the MLM without sacrificing the robustness against SNR mismatch. Unfortunately, its performance is still inferior to that of the LM algorithm with exact SNR knowledge over the class of high-loss channels. As our second contribution, a switching turbo equalization scheme, which switches between the SMLM and LM schemes, is proposed to practically close the performance gap. It is based on a novel way to estimate the SNR from the reliability values of the extrinsic information of the SMLM algorithm.
A SAR Back Projection Autofocusing Algorithm Based on Legendre Approximation
Gao Yang
2014-06-01
Full Text Available The Back Projection (BP algorithm is a very important time-domain methodology for Synthetic Aperture Radar (SAR imaging. However, conventional autofocus techniques are based on frequency-domain imaging algorithms, and can not be directly applied to BP imagery for error phase estimation. In this paper, an autofocus algorithm for BP imagery is proposed. The algorithm takes image sharpness as an objective function, and employs the coordinate descent optimization scheme to obtain the optimum phase-corrected variables by iterations. In the implementation, with a Legendre approximation of the objective function, the optimal phase estimation can be found analytically for each parameter within an iteration, avoiding computationally expensive line-search procedures. The experimental results with both simulated and measured data confirm the accuracy and effectiveness of the proposed algorithm.
SIMULATED ANNEALING BASED POLYNOMIAL TIME QOS ROUTING ALGORITHM FOR MANETS
Liu Lianggui; Feng Guangzeng
2006-01-01
Multi-constrained Quality-of-Service (QoS) routing is a big challenge for Mobile Ad hoc Networks (MANETs) where the topology may change constantly. In this paper a novel QoS Routing Algorithm based on Simulated Annealing (SA_RA) is proposed. This algorithm first uses an energy function to translate multiple QoS weights into a single mixed metric and then seeks to find a feasible path by simulated annealing. The paper outlines simulated annealing algorithm and analyzes the problems met when we apply it to Qos Routing (QoSR) in MANETs. Theoretical analysis and experiment results demonstrate that the proposed method is an effective approximation algorithms showing better performance than the other pertinent algorithm in seeking the (approximate) optimal configuration within a period of polynomial time.
Distortion Parameters Analysis Method Based on Improved Filtering Algorithm
ZHANG Shutuan
2013-10-01
Full Text Available In order to realize the accurate distortion parameters test of aircraft power supply system, and satisfy the requirement of corresponding equipment in the aircraft, the novel power parameters test system based on improved filtering algorithm is introduced in this paper. The hardware of the test system has the characters of s portable and high-speed data acquisition and processing, and the software parts utilize the software Labwindows/CVI as exploitation software, and adopt the pre-processing technique and adding filtering algorithm. Compare with the traditional filtering algorithm, the test system adopted improved filtering algorithm can help to increase the test accuracy. The application shows that the test system with improved filtering algorithm can realize the accurate test results, and reach to the design requirements.
Earth Observation Satellites Scheduling Based on Decomposition Optimization Algorithm
Feng Yao
2010-11-01
Full Text Available A decomposition-based optimization algorithm was proposed for solving Earth Observation Satellites scheduling problem. The problem was decomposed into task assignment main problem and single satellite scheduling sub-problem. In task assignment phase, the tasks were allocated to the satellites, and each satellite would schedule the task respectively in single satellite scheduling phase. We adopted an adaptive ant colony optimization algorithm to search the optimal task assignment scheme. Adaptive parameter adjusting strategy and pheromone trail smoothing strategy were introduced to balance the exploration and the exploitation of search process. A heuristic algorithm and a very fast simulated annealing algorithm were proposed to solve the single satellite scheduling problem. The task assignment scheme was valued by integrating the observation scheduling result of multiple satellites. The result was responded to the ant colony optimization algorithm, which can guide the search process of ant colony optimization. Computation results showed that the approach was effective to the satellites observation scheduling problem.
Mobile robot dynamic path planning based on improved genetic algorithm
Wang, Yong; Zhou, Heng; Wang, Ying
2017-08-01
In dynamic unknown environment, the dynamic path planning of mobile robots is a difficult problem. In this paper, a dynamic path planning method based on genetic algorithm is proposed, and a reward value model is designed to estimate the probability of dynamic obstacles on the path, and the reward value function is applied to the genetic algorithm. Unique coding techniques reduce the computational complexity of the algorithm. The fitness function of the genetic algorithm fully considers three factors: the security of the path, the shortest distance of the path and the reward value of the path. The simulation results show that the proposed genetic algorithm is efficient in all kinds of complex dynamic environments.
Multiparty Quantum Key Agreement Based on Quantum Search Algorithm.
Cao, Hao; Ma, Wenping
2017-03-23
Quantum key agreement is an important topic that the shared key must be negotiated equally by all participants, and any nontrivial subset of participants cannot fully determine the shared key. To date, the embed modes of subkey in all the previously proposed quantum key agreement protocols are based on either BB84 or entangled states. The research of the quantum key agreement protocol based on quantum search algorithms is still blank. In this paper, on the basis of investigating the properties of quantum search algorithms, we propose the first quantum key agreement protocol whose embed mode of subkey is based on a quantum search algorithm known as Grover's algorithm. A novel example of protocols with 5 - party is presented. The efficiency analysis shows that our protocol is prior to existing MQKA protocols. Furthermore it is secure against both external attack and internal attacks.
A New Generalized Similarity-Based Topic Distillation Algorithm
ZHOU Hongfang; DANG Xiaohui
2007-01-01
The procedure of hypertext induced topic search based on a semantic relation model is analyzed, and the reason for the topic drift of HITS algorithm was found to prove that Web pages are projected to a wrong latent semantic basis. A new concept-generalized similarity is introduced and, based on this, a new topic distillation algorithm GSTDA(generalized similarity based topic distillation algorithm) was presented to improve the quality of topic distillation. GSTDA was applied not only to avoid the topic drift, but also to explore relative topics to user query. The experimental results on 10 queries show that GSTDA reduces topic drift rate by 10% to 58% compared to that of HITS(hypertext induced topic search) algorithm, and discovers several relative topics to queries that have multiple meanings.
Fingerprint Image Segmentation Algorithm Based on Contourlet Transform Technology
Guanghua Zhang
2016-09-01
Full Text Available This paper briefly introduces two classic algorithms for fingerprint image processing, which include the soft threshold denoise algorithm of wavelet domain based on wavelet domain and the fingerprint image enhancement algorithm based on Gabor function. Contourlet transform has good texture sensitivity and can be used for the segmentation enforcement of the fingerprint image. The method proposed in this paper has attained the final fingerprint segmentation image through utilizing a modified denoising for a high-frequency coefficient after Contourlet decomposition, highlighting the fingerprint ridge line through modulus maxima detection and finally connecting the broken fingerprint line using a value filter in direction. It can attain richer direction information than the method based on wavelet transform and Gabor function and can make the positioning of detailed features more accurate. However, its ridge should be more coherent. Experiments have shown that this algorithm is obviously superior in fingerprint features detection.
A New Augmentation Based Algorithm for Extracting Maximal Chordal Subgraphs.
Bhowmick, Sanjukta; Chen, Tzu-Yi; Halappanavar, Mahantesh
2015-02-01
A graph is chordal if every cycle of length greater than three contains an edge between non-adjacent vertices. Chordal graphs are of interest both theoretically, since they admit polynomial time solutions to a range of NP-hard graph problems, and practically, since they arise in many applications including sparse linear algebra, computer vision, and computational biology. A maximal chordal subgraph is a chordal subgraph that is not a proper subgraph of any other chordal subgraph. Existing algorithms for computing maximal chordal subgraphs depend on dynamically ordering the vertices, which is an inherently sequential process and therefore limits the algorithms' parallelizability. In this paper we explore techniques to develop a scalable parallel algorithm for extracting a maximal chordal subgraph. We demonstrate that an earlier attempt at developing a parallel algorithm may induce a non-optimal vertex ordering and is therefore not guaranteed to terminate with a maximal chordal subgraph. We then give a new algorithm that first computes and then repeatedly augments a spanning chordal subgraph. After proving that the algorithm terminates with a maximal chordal subgraph, we then demonstrate that this algorithm is more amenable to parallelization and that the parallel version also terminates with a maximal chordal subgraph. That said, the complexity of the new algorithm is higher than that of the previous parallel algorithm, although the earlier algorithm computes a chordal subgraph which is not guaranteed to be maximal. We experimented with our augmentation-based algorithm on both synthetic and real-world graphs. We provide scalability results and also explore the effect of different choices for the initial spanning chordal subgraph on both the running time and on the number of edges in the maximal chordal subgraph.
The Algorithm for Rule-base Refinement on Fuzzy Set
LI Feng; WU Cui-hong; DING Xiang-wu
2006-01-01
In the course of running an artificial intelligent system many redundant rules are often produced. To refine the knowledge base, viz. to remove the redundant rules, can accelerate the reasoning and shrink the rule base. The purpose of the paper is to present the thinking on the topic and design the algorithm to remove the redundant rules from the rule base.The "abstraction" of "state variable", redundant rules and the least rule base are discussed in the paper. The algorithm on refining knowledge base is also presented.
Adaptive bad pixel correction algorithm for IRFPA based on PCNN
Leng, Hanbing; Zhou, Zuofeng; Cao, Jianzhong; Yi, Bo; Yan, Aqi; Zhang, Jian
2013-10-01
Bad pixels and response non-uniformity are the primary obstacles when IRFPA is used in different thermal imaging systems. The bad pixels of IRFPA include fixed bad pixels and random bad pixels. The former is caused by material or manufacture defect and their positions are always fixed, the latter is caused by temperature drift and their positions are always changing. Traditional radiometric calibration-based bad pixel detection and compensation algorithm is only valid to the fixed bad pixels. Scene-based bad pixel correction algorithm is the effective way to eliminate these two kinds of bad pixels. Currently, the most used scene-based bad pixel correction algorithm is based on adaptive median filter (AMF). In this algorithm, bad pixels are regarded as image noise and then be replaced by filtered value. However, missed correction and false correction often happens when AMF is used to handle complex infrared scenes. To solve this problem, a new adaptive bad pixel correction algorithm based on pulse coupled neural networks (PCNN) is proposed. Potential bad pixels are detected by PCNN in the first step, then image sequences are used periodically to confirm the real bad pixels and exclude the false one, finally bad pixels are replaced by the filtered result. With the real infrared images obtained from a camera, the experiment results show the effectiveness of the proposed algorithm.
A novel algorithm for computer based assessment
2012-01-01
Student learning outcomes have been evaluated through graded assignments and tests by most paper-based assessment systems. But computer based assessments has the opportunity to improve the efficiency of assessments process. The use of internet is also made possible
Cooperation-based Ant Colony Algorithm in WSN
Jianbin Xue
2013-04-01
Full Text Available This paper proposed a routing algorithm based on ant colony algorithm. The traditional ant colony algorithm updates pheromone according to the path length, to get the shortest path from the initial node to destination node. But MIMO system is different from the SISO system. The distance is farther but the energy is not bigger. Similarly, the closer the distance, the smaller the energy is not necessarily. So need to select the path according to the energy consumption of the path. This paper is based on the energy consumption to update the pheromone which from the cluster head node to the next hop node. Then, can find a path which the communication energy consumption is least. This algorithm can save more energy consumption of the network. The simulation results of MATLAB show that the path chosen by the algorithm is better than the simple ant colony algorithm, and the algorithm can save the network energy consumption better and can prolong the life cycle of the network.
A Flocking Based algorithm for Document Clustering Analysis
Cui, Xiaohui [ORNL; Gao, Jinzhu [ORNL; Potok, Thomas E [ORNL
2006-01-01
Social animals or insects in nature often exhibit a form of emergent collective behavior known as flocking. In this paper, we present a novel Flocking based approach for document clustering analysis. Our Flocking clustering algorithm uses stochastic and heuristic principles discovered from observing bird flocks or fish schools. Unlike other partition clustering algorithm such as K-means, the Flocking based algorithm does not require initial partitional seeds. The algorithm generates a clustering of a given set of data through the embedding of the high-dimensional data items on a two-dimensional grid for easy clustering result retrieval and visualization. Inspired by the self-organized behavior of bird flocks, we represent each document object with a flock boid. The simple local rules followed by each flock boid result in the entire document flock generating complex global behaviors, which eventually result in a clustering of the documents. We evaluate the efficiency of our algorithm with both a synthetic dataset and a real document collection that includes 100 news articles collected from the Internet. Our results show that the Flocking clustering algorithm achieves better performance compared to the K- means and the Ant clustering algorithm for real document clustering.
Phase shift extraction algorithm based on Euclidean matrix norm.
Deng, Jian; Wang, Hankun; Zhang, Desi; Zhong, Liyun; Fan, Jinping; Lu, Xiaoxu
2013-05-01
In this Letter, the character of Euclidean matrix norm (EMN) of the intensity difference between phase-shifting interferograms, which changes in sinusoidal form with the phase shifts, is presented. Based on this character, an EMN phase shift extraction algorithm is proposed. Both the simulation calculation and experimental research show that the phase shifts with high precision can be determined with the proposed EMN algorithm easily. Importantly, the proposed EMN algorithm will supply a powerful tool for the rapid calibration of the phase shifts.
Restart-Based Genetic Algorithm for the Quadratic Assignment Problem
Misevicius, Alfonsas
The power of genetic algorithms (GAs) has been demonstrated for various domains of the computer science, including combinatorial optimization. In this paper, we propose a new conceptual modification of the genetic algorithm entitled a "restart-based genetic algorithm" (RGA). An effective implementation of RGA for a well-known combinatorial optimization problem, the quadratic assignment problem (QAP), is discussed. The results obtained from the computational experiments on the QAP instances from the publicly available library QAPLIB show excellent performance of RGA. This is especially true for the real-life like QAPs.
Manipulator Neural Network Control Based on Fuzzy Genetic Algorithm
无
2001-01-01
The three-layer forward neural networks are used to establish the inverse kinem a tics models of robot manipulators. The fuzzy genetic algorithm based on the line ar scaling of the fitness value is presented to update the weights of neural net works. To increase the search speed of the algorithm, the crossover probability and the mutation probability are adjusted through fuzzy control and the fitness is modified by the linear scaling method in FGA. Simulations show that the propo sed method improves considerably the precision of the inverse kinematics solutio ns for robot manipulators and guarantees a rapid global convergence and overcome s the drawbacks of SGA and the BP algorithm.
A novel image encryption algorithm based on DNA subsequence operation.
Zhang, Qiang; Xue, Xianglian; Wei, Xiaopeng
2012-01-01
We present a novel image encryption algorithm based on DNA subsequence operation. Different from the traditional DNA encryption methods, our algorithm does not use complex biological operation but just uses the idea of DNA subsequence operations (such as elongation operation, truncation operation, deletion operation, etc.) combining with the logistic chaotic map to scramble the location and the value of pixel points from the image. The experimental results and security analysis show that the proposed algorithm is easy to be implemented, can get good encryption effect, has a wide secret key's space, strong sensitivity to secret key, and has the abilities of resisting exhaustive attack and statistic attack.
A motion retargeting algorithm based on model simplification
无
2005-01-01
A new motion retargeting algorithm is presented, which adapts the motion capture data to a new character. To make the resulting motion realistic, the physically-based optimization method is adopted. However, the optimization process is difficult to converge to the optimal value because of high complexity of the physical human model. In order to address this problem, an appropriate simplified model automatically determined by a motion analysis technique is utilized, and then motion retargeting with this simplified model as an intermediate agent is implemented. The entire motion retargeting algorithm involves three steps of nonlinearly constrained optimization: forward retargeting, motion scaling and inverse retargeting. Experimental results show the validity of this algorithm.
Quantum Image Encryption Algorithm Based on Quantum Image XOR Operations
Gong, Li-Hua; He, Xiang-Tao; Cheng, Shan; Hua, Tian-Xiang; Zhou, Nan-Run
2016-07-01
A novel encryption algorithm for quantum images based on quantum image XOR operations is designed. The quantum image XOR operations are designed by using the hyper-chaotic sequences generated with the Chen's hyper-chaotic system to control the control-NOT operation, which is used to encode gray-level information. The initial conditions of the Chen's hyper-chaotic system are the keys, which guarantee the security of the proposed quantum image encryption algorithm. Numerical simulations and theoretical analyses demonstrate that the proposed quantum image encryption algorithm has larger key space, higher key sensitivity, stronger resistance of statistical analysis and lower computational complexity than its classical counterparts.
A Secure Watermarking Algorithm Based on Coupled Map Lattice
YI Xiang; WANG Wei-ran
2005-01-01
Based on the nonlinear theory, a secure watermarking algorithm using wavelet transform and coupled map lattice is presented. The chaos is sensitive to initial conditions and has a good non-relevant correlation property, but the finite precision effect limits its application in practical digital watermarking system. To overcome the undesirable short period of chaos mapping and improve the security level of watermarking, the hyper-chaotic sequence is adopted in this algorithm. The watermark is mixed with the hyper-chaotic sequence and embedded in the wavelet domain of the host image. Experimental results and analysis are given to demonstrate that the proposed watermarking algorithm is transparent, robust and secure.
A Novel Image Encryption Algorithm Based on DNA Subsequence Operation
Zhang, Qiang; Xue, Xianglian; Wei, Xiaopeng
2012-01-01
We present a novel image encryption algorithm based on DNA subsequence operation. Different from the traditional DNA encryption methods, our algorithm does not use complex biological operation but just uses the idea of DNA subsequence operations (such as elongation operation, truncation operation, deletion operation, etc.) combining with the logistic chaotic map to scramble the location and the value of pixel points from the image. The experimental results and security analysis show that the proposed algorithm is easy to be implemented, can get good encryption effect, has a wide secret key's space, strong sensitivity to secret key, and has the abilities of resisting exhaustive attack and statistic attack. PMID:23093912
A Novel Image Encryption Algorithm Based on DNA Subsequence Operation
Qiang Zhang
2012-01-01
Full Text Available We present a novel image encryption algorithm based on DNA subsequence operation. Different from the traditional DNA encryption methods, our algorithm does not use complex biological operation but just uses the idea of DNA subsequence operations (such as elongation operation, truncation operation, deletion operation, etc. combining with the logistic chaotic map to scramble the location and the value of pixel points from the image. The experimental results and security analysis show that the proposed algorithm is easy to be implemented, can get good encryption effect, has a wide secret key's space, strong sensitivity to secret key, and has the abilities of resisting exhaustive attack and statistic attack.
Proposal of Tabu Search Algorithm Based on Cuckoo Search
Ahmed T. Sadiq Al-Obaidi
2014-03-01
Full Text Available This paper presents a new version of Tabu Search (TS based on Cuckoo Search (CS called (Tabu-Cuckoo Search TCS to reduce the effect of the TS problems. The proposed algorithm provides a more diversity to candidate solutions of TS. Two case studies have been solved using the proposed algorithm, 4-Color Map and Traveling Salesman Problem. The proposed algorithm gives a good result compare with the original, the iteration numbers are less and the local minimum or non-optimal solutions are less.
Heuristic-based scheduling algorithm for high level synthesis
Mohamed, Gulam; Tan, Han-Ngee; Chng, Chew-Lye
1992-01-01
A new scheduling algorithm is proposed which uses a combination of a resource utilization chart, a heuristic algorithm to estimate the minimum number of hardware units based on operator mobilities, and a list-scheduling technique to achieve fast and near optimal schedules. The schedule time of this algorithm is almost independent of the length of mobilities of operators as can be seen from the benchmark example (fifth order digital elliptical wave filter) presented when the cycle time was increased from 17 to 18 and then to 21 cycles. It is implemented in C on a SUN3/60 workstation.
A Learning Algorithm based on High School Teaching Wisdom
Philip, Ninan Sajeeth
2010-01-01
A learning algorithm based on primary school teaching and learning is presented. The methodology is to continuously evaluate a student and to give them training on the examples for which they repeatedly fail, until, they can correctly answer all types of questions. This incremental learning procedure produces better learning curves by demanding the student to optimally dedicate their learning time on the failed examples. When used in machine learning, the algorithm is found to train a machine on a data with maximum variance in the feature space so that the generalization ability of the network improves. The algorithm has interesting applications in data mining, model evaluations and rare objects discovery.
Rate control algorithm based on frame complexity estimation for MVC
Yan, Tao; An, Ping; Shen, Liquan; Zhang, Zhaoyang
2010-07-01
Rate control has not been well studied for multi-view video coding (MVC). In this paper, we propose an efficient rate control algorithm for MVC by improving the quadratic rate-distortion (R-D) model, which reasonably allocate bit-rate among views based on correlation analysis. The proposed algorithm consists of four levels for rate bits control more accurately, of which the frame layer allocates bits according to frame complexity and temporal activity. Extensive experiments show that the proposed algorithm can efficiently implement bit allocation and rate control according to coding parameters.
FAST UPDATE ALGORITHM FOR TCAM-BASED ROUTING LOOKUPS
王志恒; 叶强; 白英彩
2002-01-01
Routing technology has been forced to evolve towards higher capacity and per-port packet processing speed. The ability to achieve high forwarding speed is due to either software or hardware technology. TCAM (Ternary Content Addressable Memory) provides a performance advantage over other software or hardware search algorithms, often resulting in an order-of-magnitude reduction of search time. But slow updates may affect the performance of TCAM-based routing lookup. So the key is to design a table management algorithm, which supports high-speed updates in TCAMs. This paper presented three table management algorithms, and then compared their performance. Finally, the optimal one after comparing was given.
Performance evaluation of a texture-based segmentation algorithm
Sadjadi, Firooz A.
1991-07-01
Texture segmentations are crucial components of many remote sensing, scene analysis, and object recognition systems. However, very little attention has been paid to the problem of performance evaluation in the numerous algorithms that have been proposed by the image understanding community. In this paper, a particular algorithm is introduced and its performance is evaluated in a systematic manner on a wide range of scene and scenarios. Both the algorithm and the methodology used in its evaluation have significance in numerous applications in the computer-based image understanding field.
Automatic Image Registration Algorithm Based on Wavelet Transform
LIU Qiong; NI Guo-qiang
2006-01-01
An automatic image registration approach based on wavelet transform is proposed. This proposed method utilizes multiscale wavelet transform to extract feature points. A coarse-to-fine feature matching method is utilized in the feature matching phase. A two-way matching method based on cross-correlation to get candidate point pairs and a fine matching based on support strength combine to form the matching algorithm. At last, based on an affine transformation model, the parameters are iteratively refined by using the least-squares estimation approach. Experimental results have verified that the proposed algorithm can realize automatic registration of various kinds of images rapidly and effectively.
Clonal Strategy Algorithm Based on the Immune Memory
Ruo-Chen Liu; Li-Cheng Jiao; Hai-Feng Du
2005-01-01
Based on the clonal selection theory and immune memory mechanism in the natural immune system, a novel artificial immune system algorithm, Clonal Strategy Algorithm based on the Immune Memory (CSAIM), is proposed in this paper. The algorithm realizes the evolution of antibody population and the evolution of memory unit at the same time, and by using clonal selection operator, the global optimal computation can be combined with the local searching. According to antibody-antibody (Ab-Ab) affinity and antibody-antigen (Ab-Ag) affinity, the algorithm can allot adaptively the scales of memory unit and antibody population. It is proved theoretically that CSAIM is convergent with probability 1. And with the computer simulations of eight benchmark functions and one instance of traveling salesman problem (TSP), it is shown that CSAIM has strong abilities in having high convergence speed, enhancing the diversity of the population and avoiding the premature convergence to some extent.
A novel iris segmentation algorithm based on small eigenvalue analysis
Harish, B. S.; Aruna Kumar, S. V.; Guru, D. S.; Ngo, Minh Ngoc
2015-12-01
In this paper, a simple and robust algorithm is proposed for iris segmentation. The proposed method consists of two steps. In first step, iris and pupil is segmented using Robust Spatial Kernel FCM (RSKFCM) algorithm. RSKFCM is based on traditional Fuzzy-c-Means (FCM) algorithm, which incorporates spatial information and uses kernel metric as distance measure. In second step, small eigenvalue transformation is applied to localize iris boundary. The transformation is based on statistical and geometrical properties of the small eigenvalue of the covariance matrix of a set of edge pixels. Extensive experimentations are carried out on standard benchmark iris dataset (viz. CASIA-IrisV4 and UBIRIS.v2). We compared our proposed method with existing iris segmentation methods. Our proposed method has the least time complexity of O(n(i+p)) . The result of the experiments emphasizes that the proposed algorithm outperforms the existing iris segmentation methods.
Image Encryption Algorithm Based on Chaotic Economic Model
S. S. Askar
2015-01-01
Full Text Available In literature, chaotic economic systems have got much attention because of their complex dynamic behaviors such as bifurcation and chaos. Recently, a few researches on the usage of these systems in cryptographic algorithms have been conducted. In this paper, a new image encryption algorithm based on a chaotic economic map is proposed. An implementation of the proposed algorithm on a plain image based on the chaotic map is performed. The obtained results show that the proposed algorithm can successfully encrypt and decrypt the images with the same security keys. The security analysis is encouraging and shows that the encrypted images have good information entropy and very low correlation coefficients and the distribution of the gray values of the encrypted image has random-like behavior.
Cryptanalysis of an image encryption algorithm based on DNA encoding
Akhavan, A.; Samsudin, A.; Akhshani, A.
2017-10-01
Recently an image encryption algorithm based on DNA encoding and the Elliptic Curve Cryptography (ECC) is proposed. This paper aims to investigate the security the DNA-based image encryption algorithm and its resistance against chosen plaintext attack. The results of the analysis demonstrate that security of the algorithm mainly relies on one static shuffling step, with a simple confusion operation. In this study, a practical plain image recovery method is proposed, and it is shown that the images encrypted with the same key could easily be recovered using the suggested cryptanalysis method with as low as two chosen plain images. Also, a strategy to improve the security of the algorithm is presented in this paper.
A Modularity Degree Based Heuristic Community Detection Algorithm
Dongming Chen
2014-01-01
Full Text Available A community in a complex network can be seen as a subgroup of nodes that are densely connected. Discovery of community structures is a basic problem of research and can be used in various areas, such as biology, computer science, and sociology. Existing community detection methods usually try to expand or collapse the nodes partitions in order to optimize a given quality function. These optimization function based methods share the same drawback of inefficiency. Here we propose a heuristic algorithm (MDBH algorithm based on network structure which employs modularity degree as a measure function. Experiments on both synthetic benchmarks and real-world networks show that our algorithm gives competitive accuracy with previous modularity optimization methods, even though it has less computational complexity. Furthermore, due to the use of modularity degree, our algorithm naturally improves the resolution limit in community detection.
Efficient mining of association rules based on gravitational search algorithm
Fariba Khademolghorani
2011-07-01
Full Text Available Association rules mining are one of the most used tools to discover relationships among attributes in a database. A lot of algorithms have been introduced for discovering these rules. These algorithms have to mine association rules in two stages separately. Most of them mine occurrence rules which are easily predictable by the users. Therefore, this paper discusses the application of gravitational search algorithm for discovering interesting association rules. This evolutionary algorithm is based on the Newtonian gravity and the laws of motion. Furthermore, contrary to the previous methods, the proposed method in this study is able to mine the best association rules without generating frequent itemsets and is independent of the minimum support and confidence values. The results of applying this method in comparison with the method of mining association rules based upon the particle swarm optimization show that our method is successful.
Algorithmic Algebraic Combinatorics and Gröbner Bases
Klin, Mikhail; Jurisic, Aleksandar
2009-01-01
This collection of tutorial and research papers introduces readers to diverse areas of modern pure and applied algebraic combinatorics and finite geometries with a special emphasis on algorithmic aspects and the use of the theory of Grobner bases. Topics covered include coherent configurations, association schemes, permutation groups, Latin squares, the Jacobian conjecture, mathematical chemistry, extremal combinatorics, coding theory, designs, etc. Special attention is paid to the description of innovative practical algorithms and their implementation in software packages such as GAP and MAGM
QRS Detection Based on an Advanced Multilevel Algorithm
Wissam Jenkal; Rachid Latif; Ahmed Toumanari; Azzedine Dliou; Oussama El B’charri; Fadel Mrabih Rabou Maoulainine
2016-01-01
This paper presents an advanced multilevel algorithm used for the QRS complex detection. This method is based on three levels. The first permits the extraction of higher peaks using an adaptive thresholding technique. The second allows the QRS region detection. The last level permits the detection of Q, R and S waves. The proposed algorithm shows interesting results compared to recently published methods. The perspective of this work is the implementation of this method on an embedded system ...
Free Search Algorithm Based Estimation in WSN Location
ZHOU Hui; LI Dan-mei; SHAO Shi-huang; XU Chen
2009-01-01
This paper proposes a novel intelligent estimation algorithm in Wireless Sensor Network nodes location based on Free Search, which converts parameter estimation to on-line optimization of nonlinear function and estimates the coordinates of senor nodes using the Free Search optimization. Compared to the least-squares estimation algorithms, the localization accuracy has been increased significantly, which has been verified by the simulation results.
T-Algorithm-Based Logic Simulation on Distributed Systems
Sundaram, S; Patnaik, LM
1992-01-01
Increase in the complexity of VLSI digital circuit it sign demands faster logic simulation techniques than those currently available. One of the ways of speeding up existing logic simulataon algorithms is by exploiting the inherent parallelism an the sequentaal versaon. In this paper, we explore the possibility of mapping a T-algoriihm based logac samulataon algorithm onto a cluster of workstation interconnected by an ethernet. The set of gates at a particular level as partitioned by the hias...
PEA: Polymorphic Encryption Algorithm based on quantum computation
Komninos, N.; Mantas, G.
2011-01-01
In this paper, a polymorphic encryption algorithm (PEA), based on basic quantum computations, is proposed for the encryption of binary bits. PEA is a symmetric key encryption algorithm that applies different combinations of quantum gates to encrypt binary bits. PEA is also polymorphic since the states of the shared secret key control the different combinations of the ciphertext. It is shown that PEA achieves perfect secrecy and is resilient to eavesdropping and Trojan horse attacks. A securit...
Sampling-based Algorithms for Optimal Motion Planning
Karaman, Sertac
2011-01-01
During the last decade, sampling-based path planning algorithms, such as Probabilistic RoadMaps (PRM) and Rapidly-exploring Random Trees (RRT), have been shown to work well in practice and possess theoretical guarantees such as probabilistic completeness. However, little effort has been devoted to the formal analysis of the quality of the solution returned by such algorithms, e.g., as a function of the number of samples. The purpose of this paper is to fill this gap, by rigorously analyzing the asymptotic behavior of the cost of the solution returned by stochastic sampling-based algorithms as the number of samples increases. A number of negative results are provided, characterizing existing algorithms, e.g., showing that, under mild technical conditions, the cost of the solution returned by broadly used sampling-based algorithms converges almost surely to a non-optimal value. The main contribution of the paper is the introduction of new algorithms, namely, PRM* and RRT*, which are provably asymptotically opti...
Face detection based on multiple kernel learning algorithm
Sun, Bo; Cao, Siming; He, Jun; Yu, Lejun
2016-09-01
Face detection is important for face localization in face or facial expression recognition, etc. The basic idea is to determine whether there is a face in an image or not, and also its location, size. It can be seen as a binary classification problem, which can be well solved by support vector machine (SVM). Though SVM has strong model generalization ability, it has some limitations, which will be deeply analyzed in the paper. To access them, we study the principle and characteristics of the Multiple Kernel Learning (MKL) and propose a MKL-based face detection algorithm. In the paper, we describe the proposed algorithm in the interdisciplinary research perspective of machine learning and image processing. After analyzing the limitation of describing a face with a single feature, we apply several ones. To fuse them well, we try different kernel functions on different feature. By MKL method, the weight of each single function is determined. Thus, we obtain the face detection model, which is the kernel of the proposed method. Experiments on the public data set and real life face images are performed. We compare the performance of the proposed algorithm with the single kernel-single feature based algorithm and multiple kernels-single feature based algorithm. The effectiveness of the proposed algorithm is illustrated. Keywords: face detection, feature fusion, SVM, MKL
List-Based Simulated Annealing Algorithm for Traveling Salesman Problem
Shi-hua Zhan
2016-01-01
Full Text Available Simulated annealing (SA algorithm is a popular intelligent optimization algorithm which has been successfully applied in many fields. Parameters’ setting is a key factor for its performance, but it is also a tedious work. To simplify parameters setting, we present a list-based simulated annealing (LBSA algorithm to solve traveling salesman problem (TSP. LBSA algorithm uses a novel list-based cooling schedule to control the decrease of temperature. Specifically, a list of temperatures is created first, and then the maximum temperature in list is used by Metropolis acceptance criterion to decide whether to accept a candidate solution. The temperature list is adapted iteratively according to the topology of the solution space of the problem. The effectiveness and the parameter sensitivity of the list-based cooling schedule are illustrated through benchmark TSP problems. The LBSA algorithm, whose performance is robust on a wide range of parameter values, shows competitive performance compared with some other state-of-the-art algorithms.
Majorization-minimization algorithms for wavelet-based image restoration.
Figueiredo, Mário A T; Bioucas-Dias, José M; Nowak, Robert D
2007-12-01
Standard formulations of image/signal deconvolution under wavelet-based priors/regularizers lead to very high-dimensional optimization problems involving the following difficulties: the non-Gaussian (heavy-tailed) wavelet priors lead to objective functions which are nonquadratic, usually nondifferentiable, and sometimes even nonconvex; the presence of the convolution operator destroys the separability which underlies the simplicity of wavelet-based denoising. This paper presents a unified view of several recently proposed algorithms for handling this class of optimization problems, placing them in a common majorization-minimization (MM) framework. One of the classes of algorithms considered (when using quadratic bounds on nondifferentiable log-priors) shares the infamous "singularity issue" (SI) of "iteratively reweighted least squares" (IRLS) algorithms: the possibility of having to handle infinite weights, which may cause both numerical and convergence issues. In this paper, we prove several new results which strongly support the claim that the SI does not compromise the usefulness of this class of algorithms. Exploiting the unified MM perspective, we introduce a new algorithm, resulting from using l1 bounds for nonconvex regularizers; the experiments confirm the superior performance of this method, when compared to the one based on quadratic majorization. Finally, an experimental comparison of the several algorithms, reveals their relative merits for different standard types of scenarios.
Fuzzy logic-based diagnostic algorithm for implantable cardioverter defibrillators.
Bárdossy, András; Blinowska, Aleksandra; Kuzmicz, Wieslaw; Ollitrault, Jacky; Lewandowski, Michał; Przybylski, Andrzej; Jaworski, Zbigniew
2014-02-01
The paper presents a diagnostic algorithm for classifying cardiac tachyarrhythmias for implantable cardioverter defibrillators (ICDs). The main aim was to develop an algorithm that could reduce the rate of occurrence of inappropriate therapies, which are often observed in existing ICDs. To achieve low energy consumption, which is a critical factor for implantable medical devices, very low computational complexity of the algorithm was crucial. The study describes and validates such an algorithm and estimates its clinical value. The algorithm was based on the heart rate variability (HRV) analysis. The input data for our algorithm were: RR-interval (I), as extracted from raw intracardiac electrogram (EGM), and in addition two other features of HRV called here onset (ONS) and instability (INST). 6 diagnostic categories were considered: ventricular fibrillation (VF), ventricular tachycardia (VT), sinus tachycardia (ST), detection artifacts and irregularities (including extrasystoles) (DAI), atrial tachyarrhythmias (ATF) and no tachycardia (i.e. normal sinus rhythm) (NT). The initial set of fuzzy rules based on the distributions of I, ONS and INST in the 6 categories was optimized by means of a software tool for automatic rule assessment using simulated annealing. A training data set with 74 EGM recordings was used during optimization, and the algorithm was validated with a validation data set with 58 EGM recordings. Real life recordings stored in defibrillator memories were used. Additionally the algorithm was tested on 2 sets of recordings from the PhysioBank databases: MIT-BIH Arrhythmia Database and MIT-BIH Supraventricular Arrhythmia Database. A custom CMOS integrated circuit implementing the diagnostic algorithm was designed in order to estimate the power consumption. A dedicated Web site, which provides public online access to the algorithm, has been created and is available for testing it. The total number of events in our training and validation sets was 132. In
A Novel Algorithm Based on 3D-MUSIC Algorithm for Localizing Near-Field Source
SHAN Zhi-yong; ZHOU Xi-lang; PEN Gen-jiang
2005-01-01
A novel 3-D MUSIC algorithm based on the classical 3D-MUSIC algorithm for the location of near-field source was presented. Under the far-field assumption of actual near-field, two algebraic relations of the location parameters between the actual near-field sources and the far-field ones were derived. With Fourier transformation and polynomial-root methods, the elevation and the azimuth of the far-field were obtained, the tracking paths can be developed, and the location parameters of the near-field source can be determined, then the more accurate results can be estimated using an optimization method. The computer simulation results p rove that the algorithm for the location of the near-fields is more accurate, effective and suitable for real-time applications.
A Color Image Edge Detection Algorithm Based on Color Difference
Zhuo, Li; Hu, Xiaochen; Jiang, Liying; Zhang, Jing
2016-12-01
Although image edge detection algorithms have been widely applied in image processing, the existing algorithms still face two important problems. On one hand, to restrain the interference of noise, smoothing filters are generally exploited in the existing algorithms, resulting in loss of significant edges. On the other hand, since the existing algorithms are sensitive to noise, many noisy edges are usually detected, which will disturb the subsequent processing. Therefore, a color image edge detection algorithm based on color difference is proposed in this paper. Firstly, a new operation called color separation is defined in this paper, which can reflect the information of color difference. Then, for the neighborhood of each pixel, color separations are calculated in four different directions to detect the edges. Experimental results on natural and synthetic images show that the proposed algorithm can remove a large number of noisy edges and be robust to the smoothing filters. Furthermore, the proposed edge detection algorithm is applied in road foreground segmentation and shadow removal, which achieves good performances.
A face recognition algorithm based on thermal and visible data
Sochenkov, Ilya; Tihonkih, Dmitrii; Vokhmintcev, Aleksandr; Melnikov, Andrey; Makovetskii, Artyom
2016-09-01
In this work we present an algorithm of fusing thermal infrared and visible imagery to identify persons. The proposed face recognition method contains several components. In particular this is rigid body image registration. The rigid registration is achieved by a modified variant of the iterative closest point (ICP) algorithm. We consider an affine transformation in three-dimensional space that preserves the angles between the lines. An algorithm of matching is inspirited by the recent results of neurophysiology of vision. Also we consider the ICP minimizing error metric stage for the case of an arbitrary affine transformation. Our face recognition algorithm also uses the localized-contouring algorithms to segment the subject's face; thermal matching based on partial least squares discriminant analysis. Thermal imagery face recognition methods are advantageous when there is no control over illumination or for detecting disguised faces. The proposed algorithm leads to good matching accuracies for different person recognition scenarios (near infrared, far infrared, thermal infrared, viewed sketch). The performance of the proposed face recognition algorithm in real indoor environments is presented and discussed.
Genetic Algorithm based PID controller for Frequency Regulation Ancillary services
Sandeep Bhongade
2010-12-01
Full Text Available In this paper, the parameters of Proportional, Integral and Derivative (PID controller for Automatic Generation Control (AGC suitable in restructured power system is tuned according to Generic Algorithms (GAs based performance indices. The key idea of the proposed method is to use the fitness function based on Area Control Error (ACE. The functioning of the proposed Genetic Algorithm based PID (GAPID controller has been demonstrated on a 75-bus Indian power system network and the results have been compared with those obtained by using Least Square Minimization method.
Application of genetic algorithm to hexagon-based motion estimation.
Kung, Chih-Ming; Cheng, Wan-Shu; Jeng, Jyh-Horng
2014-01-01
With the improvement of science and technology, the development of the network, and the exploitation of the HDTV, the demands of audio and video become more and more important. Depending on the video coding technology would be the solution for achieving these requirements. Motion estimation, which removes the redundancy in video frames, plays an important role in the video coding. Therefore, many experts devote themselves to the issues. The existing fast algorithms rely on the assumption that the matching error decreases monotonically as the searched point moves closer to the global optimum. However, genetic algorithm is not fundamentally limited to this restriction. The character would help the proposed scheme to search the mean square error closer to the algorithm of full search than those fast algorithms. The aim of this paper is to propose a new technique which focuses on combing the hexagon-based search algorithm, which is faster than diamond search, and genetic algorithm. Experiments are performed to demonstrate the encoding speed and accuracy of hexagon-based search pattern method and proposed method.
Digital Image Encryption Algorithm Design Based on Genetic Hyperchaos
Jian Wang
2016-01-01
Full Text Available In view of the present chaotic image encryption algorithm based on scrambling (diffusion is vulnerable to choosing plaintext (ciphertext attack in the process of pixel position scrambling, we put forward a image encryption algorithm based on genetic super chaotic system. The algorithm, by introducing clear feedback to the process of scrambling, makes the scrambling effect related to the initial chaos sequence and the clear text itself; it has realized the image features and the organic fusion of encryption algorithm. By introduction in the process of diffusion to encrypt plaintext feedback mechanism, it improves sensitivity of plaintext, algorithm selection plaintext, and ciphertext attack resistance. At the same time, it also makes full use of the characteristics of image information. Finally, experimental simulation and theoretical analysis show that our proposed algorithm can not only effectively resist plaintext (ciphertext attack, statistical attack, and information entropy attack but also effectively improve the efficiency of image encryption, which is a relatively secure and effective way of image communication.
Face Recognition Algorithms Based on Transformed Shape Features
Sambhunath Biswas
2012-05-01
Full Text Available Human face recognition is, indeed, a challenging task, especially under illumination and pose variations. We examine in the present paper effectiveness of two simple algorithms using coiflet packet and Radon transforms to recognize human faces from some databases of still gray level images, under the environment of illumination and pose variations. Both the algorithms convert 2-D gray level training face images into their respective depth maps or physical shape which are subsequently transformed by Coiflet packet and Radon transforms to compute energy for feature extraction. Experiments show that such transformed shape features are robust to illumination and pose variations. With the features extracted, training classes are optimally separated through linear discriminant analysis (LDA, while classification for test face images is made through a k-NN classifier, based on L1 norm and Mahalanobis distance measures. Proposed algorithms are then tested on face images that differ in illumination,expression or pose separately, obtained from three databases,namely, ORL, Yale and Essex-Grimace databases. Results, so obtained, are compared with two different existing algorithms.Performance using Daubechies wavelets is also examined. It is seen that the proposed Coiflet packet and Radon transform based algorithms have significant performance, especially under different illumination conditions and pose variation. Comparison shows the proposed algorithms are superior.
Research on Quantum Searching Algorithms Based on Phase Shifts
ZHONG Pu-Cha; BAO Wan-Su
2008-01-01
@@ One iterative in Grover's original quantum search algorithm consists of two Hadamard-Walsh transformations, a selective amplitude inversion and a diffusion amplitude inversion. We concentrate on the relation among the probability of success of the algorithm, the phase shifts, the number of target items and the number of iterations via replacing the two amplitude inversions by phase shifts of an arbitrary φ = ψ(0 ≤φ, ψ≤ 2π). Then, according to the relation we find out the optimal phase shifts when the number of iterations is given. We present a new quantum search algorithm based on the optimal phase shifts of 1.018 after 0.5π /√M/N iterations. The new algorithm can obtain either a single target item or multiple target items in the search space with the probability of success at least 93.43%.
Electronic Commerce Logistics Network Optimization Based on Swarm Intelligent Algorithm
Yabing Jiao
2013-09-01
Full Text Available This article establish an efficient electronic commerce logistics operation system to reduce distribution costs and build a logistics network operation model based on around the B2C electronic commerce enterprise logistics network operation system. B2C electronic commerce transactions features in the enterprise network platform. To solve the NP-hard problem this article use hybrid ant colony algorithm, particle swarm algorithm and group swarm intelligence algorithm to get a best solution. According to the intelligent algorithm, design of electronic commerce logistics network optimization system, enter the national 22 electronic commerce logistics network for validation. Through the experiment to verify the optimized logistics cost greatly decreased. This research can help B2C electronic commerce enterprise logistics network to optimize decision-making under the premise of ensuring the interests of consumers and service levels also can be an effective way for enterprises to improve the efficiency of logistics services and reduce operation costs
PCNN document segmentation method based on bacterial foraging optimization algorithm
Liao, Yanping; Zhang, Peng; Guo, Qiang; Wan, Jian
2014-04-01
Pulse Coupled Neural Network(PCNN) is widely used in the field of image processing, but it is a difficult task to define the relative parameters properly in the research of the applications of PCNN. So far the determination of parameters of its model needs a lot of experiments. To deal with the above problem, a document segmentation based on the improved PCNN is proposed. It uses the maximum entropy function as the fitness function of bacterial foraging optimization algorithm, adopts bacterial foraging optimization algorithm to search the optimal parameters, and eliminates the trouble of manually set the experiment parameters. Experimental results show that the proposed algorithm can effectively complete document segmentation. And result of the segmentation is better than the contrast algorithms.
A layer reduction based community detection algorithm on multiplex networks
Wang, Xiaodong; Liu, Jing
2017-04-01
Detecting hidden communities is important for the analysis of complex networks. However, many algorithms have been designed for single layer networks (SLNs) while just a few approaches have been designed for multiplex networks (MNs). In this paper, we propose an algorithm based on layer reduction for detecting communities on MNs, which is termed as LRCD-MNs. First, we improve a layer reduction algorithm termed as neighaggre to combine similar layers and keep others separated. Then, we use neighaggre to find the community structure hidden in MNs. Experiments on real-life networks show that neighaggre can obtain higher relative entropy than the other algorithm. Moreover, we apply LRCD-MNs on some real-life and synthetic multiplex networks and the results demonstrate that, although LRCD-MNs does not have the advantage in terms of modularity, it can obtain higher values of surprise, which is used to evaluate the quality of partitions of a network.
Knowledge Template Based Multi-perspective Car Recognition Algorithm
Bo Cai
2010-12-01
Full Text Available In order to solve the problem due to the vehicle-oriented society such as traffic jam or traffic accident, intelligent transportation system(ITS is raised and become scientist’s research focus, with the purpose of giving people better and safer driving condition and assistance. The core of intelligent transport system is the vehicle recognition and detection, and it’s the prerequisites for other related problems. Many existing vehicle recognition algorithms are aiming at one specific direction perspective, mostly front/back and side view. To make the algorithm more robust, our paper raised a vehicle recognition algorithm for oblique vehicles while also do research on front/back and side ones. The algorithm is designed based on the common knowledge of the car, such as shape, structure and so on. The experimental results of many car images show that our method has fine accuracy in car recognition.
Meteosat Images Encryption based on AES and RSA Algorithms
Boukhatem Mohammed Belkaid
2015-06-01
Full Text Available Satellite image Security is playing a vital role in the field of communication system and Internet. This work is interested in securing transmission of Meteosat images on the Internet, in public or local networks. To enhance the security of Meteosat transmission in network communication, a hybrid encryption algorithm based on Advanced Encryption Standard (AES and Rivest Shamir Adleman (RSA algorithms is proposed. AES algorithm is used for data transmission because of its higher efficiency in block encryption and RSA algorithm is used for the encryption of the key of the AES because of its management advantages in key cipher. Our encryption system generates a unique password every new session of encryption. Cryptanalysis and various experiments have been carried out and the results were reported in this paper, which demonstrate the feasibility and flexibility of the proposed scheme.
Target Image Matching Algorithm Based on Binocular CCD Ranging
Dongming Li
2014-01-01
Full Text Available This paper proposed target image in a subpixel level matching algorithm for binocular CCD ranging, which is based on the principle of binocular CCD ranging. In the paper, firstly, we introduced the ranging principle of the binocular ranging system and deduced a binocular parallax formula. Secondly, we deduced the algorithm which was named improved cross-correlation matching algorithm and cubic surface fitting algorithm for target images matched, and it could achieve a subpixel level matching for binocular CCD ranging images. Lastly, through experiment we have analyzed and verified the actual CCD ranging images, then analyzed the errors of the experimental results and corrected the formula of calculating system errors. Experimental results showed that the actual measurement accuracy of a target within 3 km was higher than 0.52%, which meet the accuracy requirements of the high precision binocular ranging.
Validation of a Bayesian-based isotope identification algorithm
Sullivan, C.J.; Stinnett, J., E-mail: stinnettjacob@gmail.com
2015-06-01
Handheld radio-isotope identifiers (RIIDs) are widely used in Homeland Security and other nuclear safety applications. However, most commercially available devices have serious problems in their ability to correctly identify isotopes. It has been reported that this flaw is largely due to the overly simplistic identification algorithms on-board the RIIDs. This paper reports on the experimental validation of a new isotope identification algorithm using a Bayesian statistics approach to identify the source while allowing for calibration drift and unknown shielding. We present here results on further testing of this algorithm and a study on the observed variation in the gamma peak energies and areas from a wavelet-based peak identification algorithm.
Facial Affect Recognition Using Regularized Discriminant Analysis-Based Algorithms
Cheng-Yuan Shih
2010-01-01
Full Text Available This paper presents a novel and effective method for facial expression recognition including happiness, disgust, fear, anger, sadness, surprise, and neutral state. The proposed method utilizes a regularized discriminant analysis-based boosting algorithm (RDAB with effective Gabor features to recognize the facial expressions. Entropy criterion is applied to select the effective Gabor feature which is a subset of informative and nonredundant Gabor features. The proposed RDAB algorithm uses RDA as a learner in the boosting algorithm. The RDA combines strengths of linear discriminant analysis (LDA and quadratic discriminant analysis (QDA. It solves the small sample size and ill-posed problems suffered from QDA and LDA through a regularization technique. Additionally, this study uses the particle swarm optimization (PSO algorithm to estimate optimal parameters in RDA. Experiment results demonstrate that our approach can accurately and robustly recognize facial expressions.
A Load Balance Routing Algorithm Based on Uneven Clustering
Liang Yuan
2013-10-01
Full Text Available Aiming at the problem of uneven load in clustering Wireless Sensor Network (WSN, a kind of load balance routing algorithm based on uneven clustering is proposed to do uneven clustering and calculate optimal number of clustering. This algorithm prevents the number of common node under some certain cluster head from being too large which leads load to be overweight to death through even node clustering. It constructs evaluation function which can better reflect residual energy distribution of nodes and at the same time constructs routing evaluation function between cluster heads which uses MATLAB to do simulation on the performance of this algorithm. Simulation result shows that the routing established by this algorithm effectively improves network’s energy balance and lengthens the life cycle of network.
Dynamic Obfuscation Algorithm based on Demand-Driven Symbolic Execution
Yubo Yang
2014-06-01
Full Text Available Dynamic code obfuscation technique increases the difficulty of dynamically reverse by the runtime confusion. Path explosion directly affects the efficiency and accuracy of dynamic symbolic analysis. Because of the defect, this paper presents a novel algorithm DDD (Demand-Driven Dynamic Obfuscation Algorithm by using the demand-driven theory of symbolic analysis. First, create a large number of invalid paths to mislead the result of symbolic analysis. Second, according to the demand-driven theory, create a specific execution path to protect the security of software. The design and implementation of the algorithm is based on the current popular and mature SMT (satisfiability model theory, and the experimental effects are tested by Z3 - the SMT solver and Pex - the symbolic execution test tools. The experimental results prove that the algorithm enhance the security of the program.
Measuring Disorientation Based on the Needleman-Wunsch Algorithm
Tolga Güyer
2015-04-01
Full Text Available This study offers a new method to measure navigation disorientation in web based systems which is powerful learning medium for distance and open education. The Needleman-Wunsch algorithm is used to measure disorientation in a more precise manner. The process combines theoretical and applied knowledge from two previously distinct research areas, disorientation and string-matching. String-matching algorithms provide a more convenient disorientation measurement than other techniques, in that they examine the similarity between an optimal path and learners’ navigation paths. The algorithm particularly takes into account the contextual similarity between partly relevant web-pages in a user’s navigation path and pages in an optimal path. This study focuses on the reasons and the required steps to use this algorithm for disorientation measurement. Examples of actual student activities and learning environment data are provided to illustrate the process.
Hybrid Collision Detection Algorithm based on Image Space
XueLi Shen
2013-07-01
Full Text Available Collision detection is an important application in the field of virtual reality, and efficiently completing collision detection has become the research focus. For the poorly real-time defect of collision detection, this paper has presented an algorithm based on the hybrid collision detection, detecting the potential collision object sets quickly with the mixed bounding volume hierarchy tree, and then using the streaming pattern collision detection algorithm to make an accurate detection. With the above methods, it can achieve the purpose of balancing load of the CPU and GPU and speeding up the detection rate. The experimental results show that compared with the classic Rapid algorithm, this algorithm can effectively improve the efficiency of collision detection.
Medical Images Watermarking Algorithm Based on Improved DCT
Yv-fan SHANG
2013-12-01
Full Text Available Targeting at the incessant securities problems of digital information management system in modern medical system, this paper presents the robust watermarking algorithm for medical images based on Arnold transformation and DCT. The algorithm first deploys the scrambling technology to encrypt the watermark information and then combines it with the visual feature vector of the image to generate a binary logic series through the hash function. The sequence as taken as keys and stored in the third party to obtain ownership of the original image. Having no need for artificial selection of a region of interest, no capacity constraint, no participation of the original medical image, such kind of watermark extracting solves security and speed problems in the watermark embedding and extracting. The simulation results also show that the algorithm is simple in operation and excellent in robustness and invisibility. In a word, it is more practical compared with other algorithms
A robust DCT domain watermarking algorithm based on chaos system
Xiao, Mingsong; Wan, Xiaoxia; Gan, Chaohua; Du, Bo
2009-10-01
Digital watermarking is a kind of technique that can be used for protecting and enforcing the intellectual property (IP) rights of the digital media like the digital images containting in the transaction copyright. There are many kinds of digital watermarking algorithms. However, existing digital watermarking algorithms are not robust enough against geometric attacks and signal processing operations. In this paper, a robust watermarking algorithm based on chaos array in DCT (discrete cosine transform)-domain for gray images is proposed. The algorithm provides an one-to-one method to extract the watermark.Experimental results have proved that this new method has high accuracy and is highly robust against geometric attacks, signal processing operations and geometric transformations. Furthermore, the one who have on idea of the key can't find the position of the watermark embedded in. As a result, the watermark not easy to be modified, so this scheme is secure and robust.
Solution of optimal power flow using evolutionary-based algorithms
This paper applies two reliable and efficient evolutionary-based methods named Shuffled Frog Leaping Algorithm ... Grey Wolf Optimizer (GWO) to solve Optimal Power Flow (OPF) problem. OPF is ..... The wolves search for the prey based on the alpha, beta, and delta positions. ..... Energy Conversion and Management, Vol.
Knowledge Automatic Indexing Based on Concept Lexicon and Segmentation Algorithm
WANG Lan-cheng; JIANG Dan; LE Jia-jin
2005-01-01
This paper is based on two existing theories about automatic indexing of thematic knowledge concept. The prohibit-word table with position information has been designed. The improved Maximum Matching-Minimum Backtracking method has been researched. Moreover it has been studied on improved indexing algorithm and application technology based on rules and thematic concept word table.
Approximation Algorithms for Model-Based Diagnosis
Feldman, A.B.
2010-01-01
Model-based diagnosis is an area of abductive inference that uses a system model, together with observations about system behavior, to isolate sets of faulty components (diagnoses) that explain the observed behavior, according to some minimality criterion. This thesis presents greedy approximation a
Approximation Algorithms for Model-Based Diagnosis
Feldman, A.B.
2010-01-01
Model-based diagnosis is an area of abductive inference that uses a system model, together with observations about system behavior, to isolate sets of faulty components (diagnoses) that explain the observed behavior, according to some minimality criterion. This thesis presents greedy approximation a
TWO NEW FCT ALGORITHMS BASED ON PRODUCT SYSTEM
Guo Zhaoli; Shi Baochang; Wang Nengchao
2001-01-01
In this paper we present a product system and give arepresentation for consine functions with the sys tem. Based on the formula two new algorithms are designed for computing the Discrete Cosine Transform. Both algorithms have regular recursive structure and good numerical stability and are easy to parallize. CLC Number：O17 Document ID：AReferences：[1]Arguello,F. and Zapata,E. L. ,Fast Cosine Transform Based on the Successive Doubling Method,Electronics Lett.,26:19,1990,1616-1618.[2]Chan,S.C. and Ho,K.L. ,Direct Methods for Computing Discrete Sinusoidal Transform,IEE Proceedings,136: 6,1990,433- 442.[3]Chan,S.C. and Ho,K.L. ,A New Two-Dimensional Fast Cosine Transform Algorithm,IEEE Trans. Signal Processing,32:2,1991,481-485.[4]Cvetkovic,Z. and Popovic,M. V.,New Fast Recursive Algorithms for the Computation of Discrete Cosine and Sine Transforms,IEEE Trans. Signal Processing,40: 8,1992,2083-2086.[5]Hou,H.S.,A Fast Recursive Algorithm for Computing the Discrete Cosine Transform,IEEE Trans. ASSP-35:10,1987,1455-1461.[6]Lee,B.G. ,A New Algorithm to Compute the Discrete Cosine Transform,IEEE Trans. ASSP,Vol. ASSP-32:6,1984,1243-1245.[7]Lee,P. and Uang,F. Y.,Restructured Recursive DCT and DST Algorithms,IEEE Trans.Signal Processing,42: 7,1994,1600- 1609.[8]Yun,D. and Lee,S.U. ,On the Fixed-Point Error Analysis of Several Fast IDCT Algorithms,IEEE Trans. Circuits and Systems- I : Analog and Digital Signal Processing,42 : 11,1995,686- 692.Manuscript Received：2000年2月20日Published：2001年9月1日
Relevance Feedback Algorithm Based on Collaborative Filtering in Image Retrieval
Yan Sun
2010-12-01
Full Text Available Content-based image retrieval is a very dynamic study field, and in this field, how to improve retrieval speed and retrieval accuracy is a hot issue. The retrieval performance can be improved when applying relevance feedback to image retrieval and introducing the participation of people to the retrieval process. However, as for many existing image retrieval methods, there are disadvantages of relevance feedback with information not being fully saved and used, and their accuracy and flexibility are relatively poor. Based on this, the collaborative filtering technology was combined with relevance feedback in this study, and an improved relevance feedback algorithm based on collaborative filtering was proposed. In the method, the collaborative filtering technology was used not only to predict the semantic relevance between images in database and the retrieval samples, but to analyze feedback log files in image retrieval, which can make the historical data of relevance feedback be fully used by image retrieval system, and further to improve the efficiency of feedback. The improved algorithm presented has been tested on the content-based image retrieval database, and the performance of the algorithm has been analyzed and compared with the existing algorithms. The experimental results showed that, compared with the traditional feedback algorithms, this method can obviously improve the efficiency of relevance feedback, and effectively promote the recall and precision of image retrieval.
Designers' Cognitive Thinking Based on Evolutionary Algorithms
Zhang Shutao; Jianning Su; Chibing Hu; Peng Wang
2013-01-01
The research on cognitive thinking is important to construct the efficient intelligent design systems. But it is difficult to describe the model of cognitive thinking with reasonable mathematical theory. Based on the analysis of design strategy and innovative thinking, we investigated the design cognitive thinking model that included the external guide thinking of "width priority - depth priority" and the internal dominated thinking of "divergent thinking - convergent thinking", built a reaso...
Quantum Behaved Particle Swarm Optimization Algorithm Based on Artificial Fish Swarm
Dong Yumin; Zhao Li
2014-01-01
Quantum behaved particle swarm algorithm is a new intelligent optimization algorithm; the algorithm has less parameters and is easily implemented. In view of the existing quantum behaved particle swarm optimization algorithm for the premature convergence problem, put forward a quantum particle swarm optimization algorithm based on artificial fish swarm. The new algorithm based on quantum behaved particle swarm algorithm, introducing the swarm and following activities, meanwhile using the a...
van der Lee, J H; Svrcek, W Y; Young, B R
2008-01-01
Model Predictive Control is a valuable tool for the process control engineer in a wide variety of applications. Because of this the structure of an MPC can vary dramatically from application to application. There have been a number of works dedicated to MPC tuning for specific cases. Since MPCs can differ significantly, this means that these tuning methods become inapplicable and a trial and error tuning approach must be used. This can be quite time consuming and can result in non-optimum tuning. In an attempt to resolve this, a generalized automated tuning algorithm for MPCs was developed. This approach is numerically based and combines a genetic algorithm with multi-objective fuzzy decision-making. The key advantages to this approach are that genetic algorithms are not problem specific and only need to be adapted to account for the number and ranges of tuning parameters for a given MPC. As well, multi-objective fuzzy decision-making can handle qualitative statements of what optimum control is, in addition to being able to use multiple inputs to determine tuning parameters that best match the desired results. This is particularly useful for multi-input, multi-output (MIMO) cases where the definition of "optimum" control is subject to the opinion of the control engineer tuning the system. A case study will be presented in order to illustrate the use of the tuning algorithm. This will include how different definitions of "optimum" control can arise, and how they are accounted for in the multi-objective decision making algorithm. The resulting tuning parameters from each of the definition sets will be compared, and in doing so show that the tuning parameters vary in order to meet each definition of optimum control, thus showing the generalized automated tuning algorithm approach for tuning MPCs is feasible.
Plagiarism Detection Based on SCAM Algorithm
Anzelmi, Daniele; Carlone, Domenico; Rizzello, Fabio
2011-01-01
Plagiarism is a complex problem and considered one of the biggest in publishing of scientific, engineering and other types of documents. Plagiarism has also increased with the widespread use of the Internet as large amount of digital data is available. Plagiarism is not just direct copy but also...... paraphrasing, rewording, adapting parts, missing references or wrong citations. This makes the problem more difficult to handle adequately. Plagiarism detection techniques are applied by making a distinction between natural and programming languages. Our proposed detection process is based on natural language...... document. Our plagiarism detection system, like many Information Retrieval systems, is evaluated with metrics of precision and recall....
Plagiarism Detection Based on SCAM Algorithm
Anzelmi, Daniele; Carlone, Domenico; Rizzello, Fabio
2011-01-01
Plagiarism is a complex problem and considered one of the biggest in publishing of scientific, engineering and other types of documents. Plagiarism has also increased with the widespread use of the Internet as large amount of digital data is available. Plagiarism is not just direct copy but also...... paraphrasing, rewording, adapting parts, missing references or wrong citations. This makes the problem more difficult to handle adequately. Plagiarism detection techniques are applied by making a distinction between natural and programming languages. Our proposed detection process is based on natural language...
Novel Frequency Hopping Sequences Generator Based on AES Algorithm
李振荣; 庄奕琪; 张博; 张超
2010-01-01
A novel frequency hopping(FH) sequences generator based on advanced encryption standard(AES) iterated block cipher is proposed for FH communication systems.The analysis shows that the FH sequences based on AES algorithm have good performance in uniformity, correlation, complexity and security.A high-speed, low-power and low-cost ASIC of FH sequences generator is implemented by optimizing the structure of S-Box and MixColumns of AES algorithm, proposing a hierarchical power management strategy, and applying ...
Optimal design of steel portal frames based on genetic algorithms
Yue CHEN; Kai HU
2008-01-01
As for the optimal design of steel portal frames, due to both the complexity of cross selections of beams and columns and the discreteness of design variables, it is difficult to obtain satisfactory results by traditional optimization. Based on a set of constraints of the Technical Specification for Light-weighted Steel Portal Frames of China, a genetic algorithm (GA) optimization program for portal frames, written in MATLAB code, was proposed in this paper. The graph user interface (GUI) is also developed for this optimal program, so that it can be used much more conveniently. Finally, some examples illustrate the effectiveness and efficiency of the genetic-algorithm-based optimal program.
A novel bit-quad-based Euler number computing algorithm
Yao, Bin; He, Lifeng; Kang, Shiying; Chao, Yuyan; Xiao ZHAO
2015-01-01
The Euler number of a binary image is an important topological property in computer vision and pattern recognition. This paper proposes a novel bit-quad-based Euler number computing algorithm. Based on graph theory and analysis on bit-quad patterns, our algorithm only needs to count two bit-quad patterns. Moreover, by use of the information obtained during processing the previous bit-quad, the average number of pixels to be checked for processing a bit-quad is only 1.75. Experimental results ...
Half-global discretization algorithm based on rough set theory
Tan Xu; Chen Yingwu
2009-01-01
It is being widely studied how to extract knowledge from a decision table based on rough set theory. The novel problem is how to discretize a decision table having continuous attribute. In order to obtain more reasonable discretization results, a discretization algorithm is proposed, which arranges half-global discretization based on the correlational coefficient of each continuous attribute while considering the uniqueness of rough set theory. When choosing heuristic information, stability is combined with rough entropy. In terms of stability, the possibility of classifying objects belonging to certain sub-interval of a given attribute into neighbor sub-intervals is minimized. By doing this, rational discrete intervals can be determined. Rough entropy is employed to decide the optimal cut-points while guaranteeing the consistency of the decision table after discretization. Thought of this algorithm is elaborated through Iris data and then some experiments by comparing outcomes of four discritized datasets are also given, which are calculated by the proposed algorithm and four other typical algorithms for discritization respectively. After that, classification rules are deduced and summarized through rough set based classifiers. Results show that the proposed discretization algorithm is able to generate optimal classification accuracy while minimizing the number of discrete intervals. It displays superiority especially when dealing with a decision table having a large attribute number.
A Novel Multiobjective Evolutionary Algorithm Based on Regression Analysis
Zhiming Song
2015-01-01
Full Text Available As is known, the Pareto set of a continuous multiobjective optimization problem with m objective functions is a piecewise continuous (m-1-dimensional manifold in the decision space under some mild conditions. However, how to utilize the regularity to design multiobjective optimization algorithms has become the research focus. In this paper, based on this regularity, a model-based multiobjective evolutionary algorithm with regression analysis (MMEA-RA is put forward to solve continuous multiobjective optimization problems with variable linkages. In the algorithm, the optimization problem is modelled as a promising area in the decision space by a probability distribution, and the centroid of the probability distribution is (m-1-dimensional piecewise continuous manifold. The least squares method is used to construct such a model. A selection strategy based on the nondominated sorting is used to choose the individuals to the next generation. The new algorithm is tested and compared with NSGA-II and RM-MEDA. The result shows that MMEA-RA outperforms RM-MEDA and NSGA-II on the test instances with variable linkages. At the same time, MMEA-RA has higher efficiency than the other two algorithms. A few shortcomings of MMEA-RA have also been identified and discussed in this paper.
A novel multiobjective evolutionary algorithm based on regression analysis.
Song, Zhiming; Wang, Maocai; Dai, Guangming; Vasile, Massimiliano
2015-01-01
As is known, the Pareto set of a continuous multiobjective optimization problem with m objective functions is a piecewise continuous (m - 1)-dimensional manifold in the decision space under some mild conditions. However, how to utilize the regularity to design multiobjective optimization algorithms has become the research focus. In this paper, based on this regularity, a model-based multiobjective evolutionary algorithm with regression analysis (MMEA-RA) is put forward to solve continuous multiobjective optimization problems with variable linkages. In the algorithm, the optimization problem is modelled as a promising area in the decision space by a probability distribution, and the centroid of the probability distribution is (m - 1)-dimensional piecewise continuous manifold. The least squares method is used to construct such a model. A selection strategy based on the nondominated sorting is used to choose the individuals to the next generation. The new algorithm is tested and compared with NSGA-II and RM-MEDA. The result shows that MMEA-RA outperforms RM-MEDA and NSGA-II on the test instances with variable linkages. At the same time, MMEA-RA has higher efficiency than the other two algorithms. A few shortcomings of MMEA-RA have also been identified and discussed in this paper.
Cosine-Based Clustering Algorithm Approach
Mohammed A. H. Lubbad
2012-02-01
Full Text Available Due to many applications need the management of spatial data; clustering large spatial databases is an important problem which tries to find the densely populated regions in the feature space to be used in data mining, knowledge discovery, or efficient information retrieval. A good clustering approach should be efficient and detect clusters of arbitrary shapes. It must be insensitive to the outliers (noise and the order of input data. In this paper Cosine Cluster is proposed based on cosine transformation, which satisfies all the above requirements. Using multi-resolution property of cosine transforms, arbitrary shape clusters can be effectively identified at different degrees of accuracy. Cosine Cluster is also approved to be highly efficient in terms of time complexity. Experimental results on very large data sets are presented, which show the efficiency and effectiveness of the proposed approach compared to other recent clustering methods.
Improved motion information-based infrared dim target tracking algorithms
Lei, Liu; Zhijian, Huang
2014-11-01
Accurate and fast tracking of infrared (IR) dim target has very important meaning for infrared precise guidance, early warning, video surveillance, etc. However, under complex backgrounds, such as clutter, varying illumination, and occlusion, the traditional tracking method often converges to a local maximum and loses the real infrared target. To cope with these problems, three improved tracking algorithm based on motion information are proposed in this paper, namely improved mean shift algorithm, improved Optical flow method and improved Particle Filter method. The basic principles and the implementing procedure of these modified algorithms for target tracking are described. Using these algorithms, the experiments on some real-life IR and color images are performed. The whole algorithm implementing processes and results are analyzed, and those algorithms for tracking targets are evaluated from the two aspects of subjective and objective. The results prove that the proposed method has satisfying tracking effectiveness and robustness. Meanwhile, it has high tracking efficiency and can be used for real-time tracking.
A Metaheuristic Algorithm Based on Chemotherapy Science: CSA
Mohammad Hassan Salmani
2017-01-01
Full Text Available Among scientific fields of study, mathematical programming has high status and its importance has led researchers to develop accurate models and effective solving approaches to addressing optimization problems. In particular, metaheuristic algorithms are approximate methods for solving optimization problems whereby good (not necessarily optimum solutions can be generated via their implementation. In this study, we propose a population-based metaheuristic algorithm according to chemotherapy method to cure cancers that mainly search the infeasible region. As in chemotherapy, Chemotherapy Science Algorithm (CSA tries to kill inappropriate solutions (cancers and bad cells of the human body; however, this would inevitably risk incidentally destroying some acceptable solutions (healthy cells. In addition, as the cycle of cancer treatment repeats over and over, the algorithm is iterated. To align chemotherapy process with the proposed algorithm, different basic terms and definitions including Infeasibility Function (IF, objective function (OF, Cell Area (CA, and Random Cells (RCs are presented in this study. In the terminology of algorithms and optimization, IF and OF are mainly applicable as criteria to compare every pair of generated solutions. Finally, we test CSA and its structure using the benchmark Traveling Salesman Problem (TSP.
A Global Path Planning Algorithm Based on Bidirectional SVGA
Taizhi Lv
2017-01-01
Full Text Available For path planning algorithms based on visibility graph, constructing a visibility graph is very time-consuming. To reduce the computing time of visibility graph construction, this paper proposes a novel global path planning algorithm, bidirectional SVGA (simultaneous visibility graph construction and path optimization by A⁎. This algorithm does not construct a visibility graph before the path optimization. However it constructs a visibility graph and searches for an optimal path at the same time. At each step, a node with the lowest estimation cost is selected to be expanded. According to the status of this node, different through lines are drawn. If this line is free-collision, it is added to the visibility graph. If not, some vertices of obstacles which are passed through by this line are added to the OPEN list for expansion. In the SVGA process, only a few visible edges which are in relation to the optimal path are drawn and the most visible edges are ignored. For taking advantage of multicore processors, this algorithm performs SVGA in parallel from both directions. By SVGA and parallel performance, this algorithm reduces the computing time and space. Simulation experiment results in different environments show that the proposed algorithm improves the time and space efficiency of path planning.
Texture orientation-based algorithm for detecting infrared maritime targets.
Wang, Bin; Dong, Lili; Zhao, Ming; Wu, Houde; Xu, Wenhai
2015-05-20
Infrared maritime target detection is a key technology for maritime target searching systems. However, in infrared maritime images (IMIs) taken under complicated sea conditions, background clutters, such as ocean waves, clouds or sea fog, usually have high intensity that can easily overwhelm the brightness of real targets, which is difficult for traditional target detection algorithms to deal with. To mitigate this problem, this paper proposes a novel target detection algorithm based on texture orientation. This algorithm first extracts suspected targets by analyzing the intersubband correlation between horizontal and vertical wavelet subbands of the original IMI on the first scale. Then the self-adaptive wavelet threshold denoising and local singularity analysis of the original IMI is combined to remove false alarms further. Experiments show that compared with traditional algorithms, this algorithm can suppress background clutter much better and realize better single-frame detection for infrared maritime targets. Besides, in order to guarantee accurate target extraction further, the pipeline-filtering algorithm is adopted to eliminate residual false alarms. The high practical value and applicability of this proposed strategy is backed strongly by experimental data acquired under different environmental conditions.
A Scheduling Algorithm Based on Petri Nets and Simulated Annealing
Rachida H. Ghoul
2007-01-01
Full Text Available This study aims at presenting a hybrid Flexible Manufacturing System "HFMS" short-term scheduling problem. Based on the art state of general scheduling algorithms, we present the meta-heuristic, we have decided to apply for a given example of HFMS. That was the study of Simulated Annealing Algorithm SA. The HFMS model based on hierarchical Petri nets, was used to represent static and dynamic behavior of the HFMS and design scheduling solutions. Hierarchical Petri nets model was regarded as being made up a set of single timed colored Petri nets models. Each single model represents one process which was composed of many operations and tasks. The complex scheduling problem was decomposed in simple sub-problems. Scheduling algorithm was applied on each sub model in order to resolve conflicts on shared production resources.
The PCNN adaptive segmentation algorithm based on visual perception
Zhao, Yanming
To solve network adaptive parameter determination problem of the pulse coupled neural network (PCNN), and improve the image segmentation results in image segmentation. The PCNN adaptive segmentation algorithm based on visual perception of information is proposed. Based on the image information of visual perception and Gabor mathematical model of Optic nerve cells receptive field, the algorithm determines adaptively the receptive field of each pixel of the image. And determines adaptively the network parameters W, M, and β of PCNN by the Gabor mathematical model, which can overcome the problem of traditional PCNN parameter determination in the field of image segmentation. Experimental results show that the proposed algorithm can improve the region connectivity and edge regularity of segmentation image. And also show the PCNN of visual perception information for segmentation image of advantage.
CBFS: high performance feature selection algorithm based on feature clearness.
Minseok Seo
Full Text Available BACKGROUND: The goal of feature selection is to select useful features and simultaneously exclude garbage features from a given dataset for classification purposes. This is expected to bring reduction of processing time and improvement of classification accuracy. METHODOLOGY: In this study, we devised a new feature selection algorithm (CBFS based on clearness of features. Feature clearness expresses separability among classes in a feature. Highly clear features contribute towards obtaining high classification accuracy. CScore is a measure to score clearness of each feature and is based on clustered samples to centroid of classes in a feature. We also suggest combining CBFS and other algorithms to improve classification accuracy. CONCLUSIONS/SIGNIFICANCE: From the experiment we confirm that CBFS is more excellent than up-to-date feature selection algorithms including FeaLect. CBFS can be applied to microarray gene selection, text categorization, and image classification.
An Improved FCM Medical Image Segmentation Algorithm Based on MMTD
Ningning Zhou
2014-01-01
Full Text Available Image segmentation plays an important role in medical image processing. Fuzzy c-means (FCM is one of the popular clustering algorithms for medical image segmentation. But FCM is highly vulnerable to noise due to not considering the spatial information in image segmentation. This paper introduces medium mathematics system which is employed to process fuzzy information for image segmentation. It establishes the medium similarity measure based on the measure of medium truth degree (MMTD and uses the correlation of the pixel and its neighbors to define the medium membership function. An improved FCM medical image segmentation algorithm based on MMTD which takes some spatial features into account is proposed in this paper. The experimental results show that the proposed algorithm is more antinoise than the standard FCM, with more certainty and less fuzziness. This will lead to its practicable and effective applications in medical image segmentation.
Genetic algorithm-based evaluation of spatial straightness error
崔长彩; 车仁生; 黄庆成; 叶东; 陈刚
2003-01-01
A genetic algorithm ( GA ) -based approach is proposed to evaluate the straightness error of spatial lines. According to the mathematical definition of spatial straightness, a verification model is established for straightness error, and the fitness function of GA is then given and the implementation techniques of the proposed algorithm is discussed in detail. The implementation techniques include real number encoding, adaptive variable range choosing, roulette wheel and elitist combination selection strategies, heuristic crossover and single point mutation schemes etc. An application example is quoted to validate the proposed algorithm. The computation result shows that the GA-based approach is a superior nonlinear parallel optimization method. The performance of the evolution population can be improved through genetic operations such as reproduction, crossover and mutation until the optimum goal of the minimum zone solution is obtained. The quality of the solution is better and the efficiency of computation is higher than other methods.
A structural comparison of measurement-based admission control algorithms
GU Yi-ran; WANG Suo-ping; WU Hai-ya
2006-01-01
Measurement-based admission control (MBAC)algorithm is designed for the relaxed real-time service. In contrast to traditional connection admission control mechanisms,the most attractive feature of MBAC algorithm is that it does not require a prior traffic model and that is very difficult for the user to come up with a tight traffic model before establishing a flow.Other advantages of MBAC include that it can achieve higher network utilization and offer quality service to users. In this article, the study of the equations in the MBAC shows that they can all be expressed in the same form. Based on the same form,some MBAC algorithms can achieve same performance only if they satisfy some conditions.
A fast image encryption algorithm based on chaotic map
Liu, Wenhao; Sun, Kehui; Zhu, Congxu
2016-09-01
Derived from Sine map and iterative chaotic map with infinite collapse (ICMIC), a new two-dimensional Sine ICMIC modulation map (2D-SIMM) is proposed based on a close-loop modulation coupling (CMC) model, and its chaotic performance is analyzed by means of phase diagram, Lyapunov exponent spectrum and complexity. It shows that this map has good ergodicity, hyperchaotic behavior, large maximum Lyapunov exponent and high complexity. Based on this map, a fast image encryption algorithm is proposed. In this algorithm, the confusion and diffusion processes are combined for one stage. Chaotic shift transform (CST) is proposed to efficiently change the image pixel positions, and the row and column substitutions are applied to scramble the pixel values simultaneously. The simulation and analysis results show that this algorithm has high security, low time complexity, and the abilities of resisting statistical analysis, differential, brute-force, known-plaintext and chosen-plaintext attacks.
An Algorithm on Generating Lattice Based on Layered Concept Lattice
Zhang Chang-sheng
2013-08-01
Full Text Available Concept lattice is an effective tool for data analysis and rule extraction, a bottleneck factor on impacting the applications of concept lattice is how to generate lattice efficiently. In this paper, an algorithm LCLG on generating lattice in batch processing based on layered concept lattice is developed, this algorithm is based on layered concept lattice, the lattice is generated downward layer by layer through concept nodes and provisional nodes in current layer; the concept nodes are found parent-child relationships upward layer by layer, then the Hasse diagram of inter-layer connection is generated; in the generated process of the lattice nodes in each layer, we do the pruning operations dynamically according to relevant properties, and delete some unnecessary nodes, such that the generating speed is improved greatly; the experimental results demonstrate that the proposed algorithm has good performance.
Node-Dependence-Based Dynamic Incentive Algorithm in Opportunistic Networks
Ruiyun Yu
2014-01-01
Full Text Available Opportunistic networks lack end-to-end paths between source nodes and destination nodes, so the communications are mainly carried out by the “store-carry-forward” strategy. Selfish behaviors of rejecting packet relay requests will severely worsen the network performance. Incentive is an efficient way to reduce selfish behaviors and hence improves the reliability and robustness of the networks. In this paper, we propose the node-dependence-based dynamic gaming incentive (NDI algorithm, which exploits the dynamic repeated gaming to motivate nodes relaying packets for other nodes. The NDI algorithm presents a mechanism of tolerating selfish behaviors of nodes. Reward and punishment methods are also designed based on the node dependence degree. Simulation results show that the NDI algorithm is effective in increasing the delivery ratio and decreasing average latency when there are a lot of selfish nodes in the opportunistic networks.
Time-Based Dynamic Trust Model Using Ant Colony Algorithm
TANG Zhuo; LU Zhengding; LI Kai
2006-01-01
The trust in distributed environment is uncertain, which is variation for various factors. This paper introduces TDTM, a model for time-based dynamic trust. Every entity in the distribute environment is endowed with a trust-vector, which figures the trust intensity between this entity and the others. The trust intensity is dynamic due to the time and the inter-operation between two entities, a method is proposed to quantify this change based on the mind of ant colony algorithm and then an algorithm for the transfer of trust relation is also proposed. Furthermore, this paper analyses the influence to the trust intensity among all entities that is aroused by the change of trust intensity between the two entities, and presents an algorithm to resolve the problem. Finally, we show the process of the trusts'change that is aroused by the time' lapse and the inter-operation through an instance.
LAHS: A novel harmony search algorithm based on learning automata
Enayatifar, Rasul; Yousefi, Moslem; Abdullah, Abdul Hanan; Darus, Amer Nordin
2013-12-01
This study presents a learning automata-based harmony search (LAHS) for unconstrained optimization of continuous problems. The harmony search (HS) algorithm performance strongly depends on the fine tuning of its parameters, including the harmony consideration rate (HMCR), pitch adjustment rate (PAR) and bandwidth (bw). Inspired by the spur-in-time responses in the musical improvisation process, learning capabilities are employed in the HS to select these parameters based on spontaneous reactions. An extensive numerical investigation is conducted on several well-known test functions, and the results are compared with the HS algorithm and its prominent variants, including the improved harmony search (IHS), global-best harmony search (GHS) and self-adaptive global-best harmony search (SGHS). The numerical results indicate that the LAHS is more efficient in finding optimum solutions and outperforms the existing HS algorithm variants.
Effective ANT based Routing Algorithm for Data Replication in MANETs
N.J. Nithya Nandhini
2013-12-01
Full Text Available In mobile ad hoc network, the nodes often move and keep on change its topology. Data packets can be forwarded from one node to another on demand. To increase the data accessibility data are replicated at nodes and made as sharable to other nodes. Assuming that all mobile host cooperative to share their memory and allow forwarding the data packets. But in reality, all nodes do not share the resources for the benefits of others. These nodes may act selfishly to share memory and to forward the data packets. This paper focuses on selfishness of mobile nodes in replica allocation and routing protocol based on Ant colony algorithm to improve the efficiency. The Ant colony algorithm is used to reduce the overhead in the mobile network, so that it is more efficient to access the data than with other routing protocols. This result shows the efficiency of ant based routing algorithm in the replication allocation.
An Initiative-Learning Algorithm Based on System Uncertainty
ZHAO Jun
2005-01-01
Initiative-learning algorithms are characterized by and hence advantageous for their independence of prior domain knowledge.Usually,their induced results could more objectively express the potential characteristics and patterns of information systems.Initiative-learning processes can be effectively conducted by system uncertainty,because uncertainty is an intrinsic common feature of and also an essential link between information systems and their induced results.Obviously,the effectiveness of such initiative-learning framework is heavily dependent on the accuracy of system uncertainty measurements.Herein,a more reasonable method for measuring system uncertainty is developed based on rough set theory and the conception of information entropy;then a new algorithm is developed on the bases of the new system uncertainty measurement and the Skowron's algorithm for mining propositional default decision rules.The proposed algorithm is typically initiative-learning.It is well adaptable to system uncertainty.As shown by simulation experiments,its comprehensive performances are much better than those of congeneric algorithms.
K. Kumaravel
2015-05-01
Full Text Available Wireless Mesh Network (WMN uses the latest technology which helps in providing end users a high quality service referred to as the Internet’s “last mile”. Also considering WMN one of the most important technologies that are employed is multicast communication. Among the several issues routing which is significantly an important issue is addressed by every WMN technologies and this is done during the process of data transmission. The IEEE 802.11s Standard entails and sets procedures which need to be followed to facilitate interconnection and thus be able to devise an appropriate WMN. There has been introduction of several protocols by many authors which are mainly devised on the basis of machine learning and artificial intelligence. Multi-path routing may be considered as one such routing method which facilitates transmission of data over several paths, proving its capabilities as a useful strategy for achieving reliability in WMN. Though, multi-path routing in any manner cannot really guarantee deterministic transmission. As here there are multiple paths available for enabling data transmission from source to destination node. The algorithm that had been employed before in the studies conducted did not take in to consideration routing metrics which include energy aware metrics that are used for path selection during transferring of data. The following study proposes use of the hybrid multipath routing algorithm while taking in to consideration routing metrics which include energy, minimal loss for efficient path selection and transferring of data. Proposed algorithm here has two phases. In the first phase prim’s algorithm has been proposed so that in networks route discovery may be possible. For the second one the Hybrid firefly algorithm which is based on harmony search has been employed for selection of the most suitable and best through proper analysis of metrics which include energy awareness and minimal loss for every path that has
A Lex-BFS-based recognition algorithm for Robinsonian matrices
Laurent, M.; Seminaroti, M.; Paschos, V.; Widmayer, P.
2015-01-01
Robinsonian matrices arise in the classical seriation problem and play an important role in many applications where unsorted similarity (or dissimilarity) information must be re- ordered. We present a new polynomial time algorithm to recognize Robinsonian matrices based on a new characterization of
A Lex-BFS-based recognition algorithm for Robinsonian matrices
M. Laurent (Monique); M. Seminaroti (Matteo); V. Paschos; P. Widmayer
2015-01-01
htmlabstractRobinsonian matrices arise in the classical seriation problem and play an important role in many applications where unsorted similarity (or dissimilarity) information must be re- ordered. We present a new polynomial time algorithm to recognize Robinsonian matrices based on a new characte
Measuring Disorientation Based on the Needleman-Wunsch Algorithm
Güyer, Tolga; Atasoy, Bilal; Somyürek, Sibel
2015-01-01
This study offers a new method to measure navigation disorientation in web based systems which is powerful learning medium for distance and open education. The Needleman-Wunsch algorithm is used to measure disorientation in a more precise manner. The process combines theoretical and applied knowledge from two previously distinct research areas,…
A Lex-BFS-based recognition algorithm for Robinsonian matrices
M. Laurent (Monique); M. Seminaroti (Matteo); V. Paschos; P. Widmayer
2015-01-01
htmlabstractRobinsonian matrices arise in the classical seriation problem and play an important role in many applications where unsorted similarity (or dissimilarity) information must be re- ordered. We present a new polynomial time algorithm to recognize Robinsonian matrices based on a new
Competition assignment problem algorithm based on Hungarian method
KONG Chao; REN Yongtai; GE Huiling; DENG Hualing
2007-01-01
Traditional Hungarian method can only solve standard assignment problems, while can not solve competition assignment problems. This article emphatically discussed the difference between standard assignment problems and competition assignment problems. The kinds of competition assignment problem algorithms based on Hungarian method and the solutions of them were studied.
Photoacoustic image reconstruction based on Bayesian compressive sensing algorithm
Mingjian Sun; Naizhang Feng; Yi Shen; Jiangang Li; Liyong Ma; Zhenghua Wu
2011-01-01
The photoacoustic tomography (PAT) method, based on compressive sensing (CS) theory, requires that,for the CS reconstruction, the desired image should have a sparse representation in a known transform domain. However, the sparsity of photoacoustic signals is destroyed because noises always exist. Therefore,the original sparse signal cannot be effectively recovered using the general reconstruction algorithm. In this study, Bayesian compressive sensing (BCS) is employed to obtain highly sparse representations of photoacoustic images based on a set of noisy CS measurements. Results of simulation demonstrate that the BCS-reconstructed image can achieve superior performance than other state-of-the-art CS-reconstruction algorithms.%@@ The photoacoustic tomography (PAT) method, based on compressive sensing (CS) theory, requires that,for the CS reconstruction, the desired image should have a sparse representation in a known transform domain.However, the sparsity of photoacoustic signals is destroyed because noises always exist.Therefore,the original sparse signal cannot be effectively recovered using the general reconstruction algorithm.In this study, Bayesian compressive sensing (BCS) is employed to obtain highly sparse representations of photoacoustic inages based on a set of noisy CS measurements.Results of simulation demonstrate that the BCS-reconstructed image can achieve superior performance than other state-of-the-art CS-reconstruction algorithms.
An Efficient 16-Bit Multiplier based on Booth Algorithm
Khan, M. Zamin Ali; Saleem, Hussain; Afzal, Shiraz; Naseem, Jawed
2012-11-01
Multipliers are key components of many high performance systems such as microprocessors, digital signal processors, etc. Optimizing the speed and area of the multiplier is major design issue which is usually conflicting constraint so that improving speed results mostly in bigger areas. A VHDL designed architecture based on booth multiplication algorithm is proposed which not only optimize speed but also efficient on energy use.
Reducing Ultrasonic Signal Noise by Algorithms based on Wavelet Thresholding
M. Kreidl
2002-01-01
Full Text Available Traditional techniques for reducing ultrasonic signal noise are based on the optimum frequency of an acoustic wave, ultrasonic probe construction and low-noise electronic circuits. This paper describes signal processing methods for noise suppression using a wavelet transform. Computer simulations of the proposed testing algorithms are presented.
A Table Based Algorithm for Minimum Directed Spanning Trees
无
2001-01-01
As far as the weighted digraph is considered, an optimal directed spanning tree algorithm called table basedalgorithm (TBA) ia proposed in the paper based on the table instead of the weighted digraph. The optimality is proved,and a numerical example is demonatrated.
A CT Image Segmentation Algorithm Based on Level Set Method
QU Jing-yi; SHI Hao-shan
2006-01-01
Level Set methods are robust and efficient numerical tools for resolving curve evolution in image segmentation. This paper proposes a new image segmentation algorithm based on Mumford-Shah module. The method is used to CT images and the experiment results demonstrate its efficiency and veracity.
ITO-based evolutionary algorithm to solve traveling salesman problem
Dong, Wenyong; Sheng, Kang; Yang, Chuanhua; Yi, Yunfei
2014-03-01
In this paper, a ITO algorithm inspired by ITO stochastic process is proposed for Traveling Salesmen Problems (TSP), so far, many meta-heuristic methods have been successfully applied to TSP, however, as a member of them, ITO needs further demonstration for TSP. So starting from designing the key operators, which include the move operator, wave operator, etc, the method based on ITO for TSP is presented, and moreover, the ITO algorithm performance under different parameter sets and the maintenance of population diversity information are also studied.
Application layer multicast routing solution based on genetic algorithms
Peng CHENG; Qiufeng WU; Qionghai DAI
2009-01-01
Application layer multicast routing is a multi-objective optimization problem.Three routing con-straints,tree's cost,tree's balance and network layer load distribution are analyzed in this paper.The three fitness functions are used to evaluate a multicast tree on the three indexes respectively and one general fitness function is generated.A novel approach based on genetic algorithms is proposed.Numerical simulations show that,compared with geometrical routing rules,the proposed algorithm improve all three indexes,especially on cost and network layer load distribution indexes.
Digital Watermarking Algorithm Based on Wavelet Transform and Neural Network
WANG Zhenfei; ZHAI Guangqun; WANG Nengchao
2006-01-01
An effective blind digital watermarking algorithm based on neural networks in the wavelet domain is presented. Firstly, the host image is decomposed through wavelet transform. The significant coefficients of wavelet are selected according to the human visual system (HVS) characteristics. Watermark bits are added to them. And then effectively cooperates neural networks to learn the characteristics of the embedded watermark related to them. Because of the learning and adaptive capabilities of neural networks, the trained neural networks almost exactly recover the watermark from the watermarked image. Experimental results and comparisons with other techniques prove the effectiveness of the new algorithm.
Image fusion based on expectation maximization algorithm and steerable pyramid
Gang Liu(刘刚); Zhongliang Jing(敬忠良); Shaoyuan Sun(孙韶媛); Jianxun Li(李建勋); Zhenhua Li(李振华); Henry Leung
2004-01-01
In this paper, a novel image fusion method based on the expectation maximization (EM) algorithm and steerable pyramid is proposed. The registered images are first decomposed by using steerable pyramid.The EM algorithm is used to fuse the image components in the low frequency band. The selection method involving the informative importance measure is applied to those in the high frequency band. The final fused image is then computed by taking the inverse transform on the composite coefficient representations.Experimental results show that the proposed method outperforms conventional image fusion methods.
Community Structure Detection Algorithm Based on the Node Belonging Degree
Jian Li
2013-07-01
Full Text Available In this paper, we propose a novel algorithm to identify communities in complex networks based on the node belonging degree. First, we give the concept of the node belonging degree, and then determine whether a node belongs to a community or not according to the belonging degree of the node with respect to the community. The experiment results of three real-world networks: a network with three communities with 19 nodes, Zachary Karate Club and network of American college football teams show that the proposed algorithm has satisfactory community structure detection.
Matrix-based, finite-difference algorithms for computational acoustics
Davis, Sanford
1990-01-01
A compact numerical algorithm is introduced for simulating multidimensional acoustic waves. The algorithm is expressed in terms of a set of matrix coefficients on a three-point spatial grid that approximates the acoustic wave equation with a discretization error of O(h exp 5). The method is based on tracking a local phase variable and its implementation suggests a convenient coordinate splitting along with natural intermediate boundary conditions. Results are presented for oblique plane waves and compared with other procedures. Preliminary computations of acoustic diffraction are also considered.
A SAR IMAGE REGISTRATION METHOD BASED ON SIFT ALGORITHM
W. Lu
2017-09-01
Full Text Available In order to improve the stability and rapidity of synthetic aperture radar (SAR images matching, an effective method was presented. Firstly, the adaptive smoothing filtering was employed for image denoising in image processing based on Wallis filtering to avoid the follow-up noise is amplified. Secondly, feature points were extracted by a simplified SIFT algorithm. Finally, the exact matching of the images was achieved with these points. Compared with the existing methods, it not only maintains the richness of features, but a-lso reduces the noise of the image. The simulation results show that the proposed algorithm can achieve better matching effect.
QRS Detection Based on an Advanced Multilevel Algorithm
Wissam Jenkal
2016-01-01
Full Text Available This paper presents an advanced multilevel algorithm used for the QRS complex detection. This method is based on three levels. The first permits the extraction of higher peaks using an adaptive thresholding technique. The second allows the QRS region detection. The last level permits the detection of Q, R and S waves. The proposed algorithm shows interesting results compared to recently published methods. The perspective of this work is the implementation of this method on an embedded system for a real time ECG monitoring system.
An Optimal Seed Based Compression Algorithm for DNA Sequences
Pamela Vinitha Eric
2016-01-01
Full Text Available This paper proposes a seed based lossless compression algorithm to compress a DNA sequence which uses a substitution method that is similar to the LempelZiv compression scheme. The proposed method exploits the repetition structures that are inherent in DNA sequences by creating an offline dictionary which contains all such repeats along with the details of mismatches. By ensuring that only promising mismatches are allowed, the method achieves a compression ratio that is at par or better than the existing lossless DNA sequence compression algorithms.
A Sumudu based algorithm for solving differential equations
Jun Zhang
2007-11-01
Full Text Available An algorithm based on Sumudu transform is developed. The algorithm can be implemented in computer algebra systems like Maple. It can be used to solve differential equations of the following form automatically without human interaction \\begin{displaymath} \\sum_{i=0}^{m} p_i(xy^{(i}(x = \\sum_{j=0}^{k}q_j(xh_j(x \\end{displaymath} where pi(x(i=0, 1, 2, ..., m and qj(x(j=0, 1, 2, ..., k are polynomials. hj(x are non-rational functions, but their Sumudu transforms are rational. m, k are nonnegative integers.
Analog Group Delay Equalizers Design Based on Evolutionary Algorithm
M. Laipert
2006-04-01
Full Text Available This paper deals with a design method of the analog all-pass filter designated for equalization of the group delay frequency response of the analog filter. This method is based on usage of evolutionary algorithm, the Differential Evolution algorithm in particular. We are able to design such equalizers to be obtained equal-ripple group delay frequency response in the pass-band of the low-pass filter. The procedure works automatically without an input estimation. The method is presented on solving practical examples.
Core Business Selection Based on Ant Colony Clustering Algorithm
Yu Lan
2014-01-01
Full Text Available Core business is the most important business to the enterprise in diversified business. In this paper, we first introduce the definition and characteristics of the core business and then descript the ant colony clustering algorithm. In order to test the effectiveness of the proposed method, Tianjin Port Logistics Development Co., Ltd. is selected as the research object. Based on the current situation of the development of the company, the core business of the company can be acquired by ant colony clustering algorithm. Thus, the results indicate that the proposed method is an effective way to determine the core business for company.
A New Algorithm for Total Variation Based Image Denoising
Yi-ping XU
2012-01-01
We propose a new algorithm for the total variation based on image denoising problem.The split Bregman method is used to convert an unconstrained minimization denoising problem to a linear system in the outer iteration.An algebraic multi-grid method is applied to solve the linear system in the inner iteration.Furthermore,Krylov subspace acceleration is adopted to improve convergence in the outer iteration.Numerical experiments demonstrate that this algorithm is efficient even for images with large signal-to-noise ratio.
Algorithms for Quantum Branching Programs Based on Fingerprinting
Ablayev, Farid; 10.4204/EPTCS.9.1
2009-01-01
In the paper we develop a method for constructing quantum algorithms for computing Boolean functions by quantum ordered read-once branching programs (quantum OBDDs). Our method is based on fingerprinting technique and representation of Boolean functions by their characteristic polynomials. We use circuit notation for branching programs for desired algorithms presentation. For several known functions our approach provides optimal QOBDDs. Namely we consider such functions as Equality, Palindrome, and Permutation Matrix Test. We also propose a generalization of our method and apply it to the Boolean variant of the Hidden Subgroup Problem.
Stellar Population Analysis of Galaxies based on Genetic Algorithms
Abdel-Fattah Attia; H.A.Ismail; I.M.Selim; A.M.Osman; I.A.Isaa; M.A.Marie; A.A.Shaker
2005-01-01
We present a new method for determining the age and relative contribution of different stellar populations in galaxies based on the genetic algorithm.We apply this method to the barred spiral galaxy NGC 3384, using CCD images in U, B, V, R and I bands. This analysis indicates that the galaxy NGC 3384 is mainly inhabited by old stellar population (age ＞ 109 yr). Some problems were encountered when numerical simulations are used for determining the contribution of different stellar populations in the integrated color of a galaxy. The results show that the proposed genetic algorithm can search efficiently through the very large space of the possible ages.
Web mining based on chaotic social evolutionary programming algorithm
无
2008-01-01
With an aim to the fact that the K-means clustering algorithm usually ends in local optimization and is hard to harvest global optimization, a new web clustering method is presented based on the chaotic social evolutionary programming (CSEP) algorithm. This method brings up the manner of that a cognitive agent inherits a paradigm in clustering to enable the cognitive agent to acquire a chaotic mutation operator in the betrayal. As proven in the experiment, this method can not only effectively increase web clustering efficiency, but it can also practically improve the precision of web clustering.
a SAR Image Registration Method Based on Sift Algorithm
Lu, W.; Yue, X.; Zhao, Y.; Han, C.
2017-09-01
In order to improve the stability and rapidity of synthetic aperture radar (SAR) images matching, an effective method was presented. Firstly, the adaptive smoothing filtering was employed for image denoising in image processing based on Wallis filtering to avoid the follow-up noise is amplified. Secondly, feature points were extracted by a simplified SIFT algorithm. Finally, the exact matching of the images was achieved with these points. Compared with the existing methods, it not only maintains the richness of features, but a-lso reduces the noise of the image. The simulation results show that the proposed algorithm can achieve better matching effect.
Design and Implementation of GPU-Based Prim's Algorithm
Wei Wang
2011-07-01
Full Text Available Minimum spanning tree is a classical problem in graph theory that plays a key role in a broad domain of applications. This paper proposes a minimum spanning tree algorithm using Prim's approach on Nvidia GPU under CUDA architecture. By using new developed GPU-based Min-Reduction data parallel primitive in the key step of the algorithm, higher efficiency is achieved. Experimental results show that we obtain about 2 times speedup on Nvidia GTX260 GPU over the CPU implementation and 3 times speedup over non-primitives GPU implementation.
CCH-based geometric algorithms for SVM and applications
Xin-jun PENG; Yi-fei WANG
2009-01-01
The support vector machine (SVM) is a novel machine learning tool in data mining. In this paper, the geometric approach based on the compressed convex hull (CCH) with a mathematical framework is introduced to solve SVM classification problems. Compared with the reduced convex hull (RCH), CCH preserves the shape of geometric solids for data sets; meanwhile, it is easy to give the necessary and sufficient condition for determining its extreme points. As practical applications of CCH, spare and probabilistic speed-up geometric algorithms are developed. Results of numerical experiments show that the proposed algorithms can reduce kernel calculations and display nice performances.
Multiple Lookup Table-Based AES Encryption Algorithm Implementation
Gong, Jin; Liu, Wenyi; Zhang, Huixin
Anew AES (Advanced Encryption Standard) encryption algorithm implementation was proposed in this paper. It is based on five lookup tables, which are generated from S-box(the substitution table in AES). The obvious advantages are reducing the code-size, improving the implementation efficiency, and helping new learners to understand the AES encryption algorithm and GF(28) multiplication which are necessary to correctly implement AES[1]. This method can be applied on processors with word length 32 or above, FPGA and others. And correspondingly we can implement it by VHDL, Verilog, VB and other languages.
Optimizing Combination of Units Commitment Based on Improved Genetic Algorithms
LAI Yifei; ZHANG Qianhua; JIA Junping
2007-01-01
GAs are general purpose optimization techniques based on principles inspired from the biological evolution using metaphors of mechanisms, such as natural selection, genetic recombination and survival of the fittest. By use of coding betterment, the dynamic changes of the mutation rate and the crossover probability, the dynamic choice of subsistence, the reservation of the optimal fitness value, a modified genetic algorithm for optimizing combination of units in thermal power plants is proposed.And through taking examples, test result are analyzed and compared with results of some different algorithms. Numerical results show available value for the unit commitment problem with examples.
Model-based Bayesian signal extraction algorithm for peripheral nerves
Eggers, Thomas E.; Dweiri, Yazan M.; McCallum, Grant A.; Durand, Dominique M.
2017-10-01
Objective. Multi-channel cuff electrodes have recently been investigated for extracting fascicular-level motor commands from mixed neural recordings. Such signals could provide volitional, intuitive control over a robotic prosthesis for amputee patients. Recent work has demonstrated success in extracting these signals in acute and chronic preparations using spatial filtering techniques. These extracted signals, however, had low signal-to-noise ratios and thus limited their utility to binary classification. In this work a new algorithm is proposed which combines previous source localization approaches to create a model based method which operates in real time. Approach. To validate this algorithm, a saline benchtop setup was created to allow the precise placement of artificial sources within a cuff and interference sources outside the cuff. The artificial source was taken from five seconds of chronic neural activity to replicate realistic recordings. The proposed algorithm, hybrid Bayesian signal extraction (HBSE), is then compared to previous algorithms, beamforming and a Bayesian spatial filtering method, on this test data. An example chronic neural recording is also analyzed with all three algorithms. Main results. The proposed algorithm improved the signal to noise and signal to interference ratio of extracted test signals two to three fold, as well as increased the correlation coefficient between the original and recovered signals by 10-20%. These improvements translated to the chronic recording example and increased the calculated bit rate between the recovered signals and the recorded motor activity. Significance. HBSE significantly outperforms previous algorithms in extracting realistic neural signals, even in the presence of external noise sources. These results demonstrate the feasibility of extracting dynamic motor signals from a multi-fascicled intact nerve trunk, which in turn could extract motor command signals from an amputee for the end goal of
Parameter Selection for Ant Colony Algorithm Based on Bacterial Foraging Algorithm
Peng Li
2016-01-01
Full Text Available The optimal performance of the ant colony algorithm (ACA mainly depends on suitable parameters; therefore, parameter selection for ACA is important. We propose a parameter selection method for ACA based on the bacterial foraging algorithm (BFA, considering the effects of coupling between different parameters. Firstly, parameters for ACA are mapped into a multidimensional space, using a chemotactic operator to ensure that each parameter group approaches the optimal value, speeding up the convergence for each parameter set. Secondly, the operation speed for optimizing the entire parameter set is accelerated using a reproduction operator. Finally, the elimination-dispersal operator is used to strengthen the global optimization of the parameters, which avoids falling into a local optimal solution. In order to validate the effectiveness of this method, the results were compared with those using a genetic algorithm (GA and a particle swarm optimization (PSO, and simulations were conducted using different grid maps for robot path planning. The results indicated that parameter selection for ACA based on BFA was the superior method, able to determine the best parameter combination rapidly, accurately, and effectively.
An ellipse detection algorithm based on edge classification
Yu, Liu; Chen, Feng; Huang, Jianming; Wei, Xiangquan
2015-12-01
In order to enhance the speed and accuracy of ellipse detection, an ellipse detection algorithm based on edge classification is proposed. Too many edge points are removed by making edge into point in serialized form and the distance constraint between the edge points. It achieves effective classification by the criteria of the angle between the edge points. And it makes the probability of randomly selecting the edge points falling on the same ellipse greatly increased. Ellipse fitting accuracy is significantly improved by the optimization of the RED algorithm. It uses Euclidean distance to measure the distance from the edge point to the elliptical boundary. Experimental results show that: it can detect ellipse well in case of edge with interference or edges blocking each other. It has higher detecting precision and less time consuming than the RED algorithm.
A Parsing Graph-based Algorithm for Ontology Mapping
WANG Zong-jiang; WANG Ying-lin; ZHANG Shen-sheng; DU Tao
2009-01-01
Ontology mapping is a critical problem for integrating the heterogeneous information sources. It can identify the elements corresponding to each other. At present, there are many ontology mapping algorithms, but most of them are bused on database schema. After analyzing the similarity and difference of ontology and schema, wepropose a parsing graph-based algorithm for ontology mapping. The ontology parsing graph (OP-graph) extends the general concept of graph, encodes logic relationship, and semantic information which the ontology contains into vertices and edges of the graph. Thus, the problem of ontology mapping is translated into a problem of finding the optimal match between the two OP-graphs. With the definition of a universal measure for comparing the entities of two ontoingies, we calculate the whole similarity between the two OP-graphs iteratively, until the optimal match is found. The results of experiments show that our algorithm is promising.
Neighborhood based Levenberg-Marquardt algorithm for neural network training.
Lera, G; Pinzolas, M
2002-01-01
Although the Levenberg-Marquardt (LM) algorithm has been extensively applied as a neural-network training method, it suffers from being very expensive, both in memory and number of operations required, when the network to be trained has a significant number of adaptive weights. In this paper, the behavior of a recently proposed variation of this algorithm is studied. This new method is based on the application of the concept of neural neighborhoods to the LM algorithm. It is shown that, by performing an LM step on a single neighborhood at each training iteration, not only significant savings in memory occupation and computing effort are obtained, but also, the overall performance of the LM method can be increased.
PSO Algorithm Based on Accumulation Effect and Mutation
Ji Wei Dong
2013-07-01
Full Text Available Particle SwarmOptimization (PSO algorithm is a new swarm intelligence optimization technique, because of its simplicity, fewerparameters and good effects, PSO has been widely used to solve various complexoptimization problems. particle swarm optimization(PSO exist the problems ofpremature and local convergence, we proposed an improved particle swarmoptimization based on aggregation effect and with mutation operator, whichdetermines whether the aggregation occurs in searching, if there is then theGaussian mutation is detected to theglobal extremum , to overcome particle swarm optimization falling into localoptimal solution defects. Testing thenew algorithm by a typical test function, the results show that , compared withthe conventional genetic algorithm(SGA, it improves the ability of globaloptimization, but also effectively avoid the premature convergence.
Applied RCM2 Algorithms Based on Statistical Methods
Fausto Pedro García Márquez; Diego J. Pedregal
2007-01-01
The main purpose of this paper is to implement a system capable of detecting faults in railway point mechanisms. This is achieved by developing an algorithm that takes advantage of three empirical criteria simultaneously capable of detecting faults from records of measurements of force against time. The system is dynamic in several respects: the base reference data is computed using all the curves free from faults as they are encountered in the experimental data; the algorithm that uses the three criteria simultaneously may be applied in on-line situations as each new data point becomes available; and recursive algorithms are applied to filter noise from the raw data in an automatic way. Encouraging results are found in practice when the system is applied to a number of experiments carried out by an industrial sponsor.
Chaos-Based Image Encryption Algorithm Using Decomposition
Xiuli Song
2013-07-01
Full Text Available The proposed chaos-based image encryption algorithm consists of four stages: decomposition, shuffle, diffusion and combination. Decomposition is that an original image is decomposed to components according to some rule. The purpose of the shuffle is to mask original organization of the pixels of the image, and the diffusion is to change their values. Combination is not necessary in the sender. To improve the efficiency, the parallel architecture is taken to process the shuffle and diffusion. To enhance the security of the algorithm, firstly, a permutation of the labels is designed. Secondly, two Logistic maps are used in diffusion stage to encrypt the components. One map encrypts the odd rows of the component and another map encrypts the even rows. Experiment results and security analysis demonstrate that the encryption algorithm not only is robust and flexible, but also can withstand common attacks such as statistical attacks and differential attacks.
Chaos-Based Encryption Algorithm for Compressed Video
袁春; 钟玉琢; 贺玉文
2003-01-01
Encryption for compressed video streams has attracted increasing attention with the exponential growth of digital multimedia delivery and consumption. However, most algorithms proposed in the literature do not effectively address the peculiarities of security and performance requirements. This paper presents a chaos-based encryption algorithm called the chaotic selective encryption of compressed video (CSECV) which exploits the characteristics of the compressed video. The encryption has three separate layers that can be selected according to the security needs of the application and the processing capability of the client computer. The chaotic pseudo-random sequence generator used to generate the key-sequence to randomize the important fields in the compressed video stream has its parameters encrypted by an asymmetric cipher and placed into the stream. The resulting stream is still a valid video stream. CSECV has significant advantages over existing algorithms for security, decryption speed, implementation flexibility, and error preservation.
Design of synthetic biological logic circuits based on evolutionary algorithm.
Chuang, Chia-Hua; Lin, Chun-Liang; Chang, Yen-Chang; Jennawasin, Tanagorn; Chen, Po-Kuei
2013-08-01
The construction of an artificial biological logic circuit using systematic strategy is recognised as one of the most important topics for the development of synthetic biology. In this study, a real-structured genetic algorithm (RSGA), which combines general advantages of the traditional real genetic algorithm with those of the structured genetic algorithm, is proposed to deal with the biological logic circuit design problem. A general model with the cis-regulatory input function and appropriate promoter activity functions is proposed to synthesise a wide variety of fundamental logic gates such as NOT, Buffer, AND, OR, NAND, NOR and XOR. The results obtained can be extended to synthesise advanced combinational and sequential logic circuits by topologically distinct connections. The resulting optimal design of these logic gates and circuits are established via the RSGA. The in silico computer-based modelling technology has been verified showing its great advantages in the purpose.
An Algorithm of Sensor Management Based on Dynamic Target Detection
LIUXianxing; ZHOULin; JINYong
2005-01-01
The probability density of stationary target is only evolved at measurement update, but the probability density of dynamic target is evolved not only at measurement update but also during measurements, this paper researches an algorithm of dynamic targets detection. Firstly, it presents the evolution of probability density at measurement update by Bayes' rule and the evolution of probability density during measurements by Fokker-Planck differential equations, respectively. Secondly, the method of obtaining information entropy by the probability density is given and sensor resources are distributed based on the evolution of information entropy viz. the maximization of information gain. Simulation results show that compared with the algorithm of serial search, this algorithm is feasible and effective when it is used to detect dynamic target.
Method of stereo matching based on genetic algorithm
Lu, Chaohui; An, Ping; Zhang, Zhaoyang
2003-09-01
A new stereo matching scheme based on image edge and genetic algorithm (GA) is presented to improve the conventional stereo matching method in this paper. In order to extract robust edge feature for stereo matching, infinite symmetric exponential filter (ISEF) is firstly applied to remove the noise of image, and nonlinear Laplace operator together with local variance of intensity are then used to detect edges. Apart from the detected edge, the polarity of edge pixels is also obtained. As an efficient search method, genetic algorithm is applied to find the best matching pair. For this purpose, some new ideas are developed for applying genetic algorithm to stereo matching. Experimental results show that the proposed methods are effective and can obtain good results.
Community Clustering Algorithm in Complex Networks Based on Microcommunity Fusion
Jin Qi
2015-01-01
Full Text Available With the further research on physical meaning and digital features of the community structure in complex networks in recent years, the improvement of effectiveness and efficiency of the community mining algorithms in complex networks has become an important subject in this area. This paper puts forward a concept of the microcommunity and gets final mining results of communities through fusing different microcommunities. This paper starts with the basic definition of the network community and applies Expansion to the microcommunity clustering which provides prerequisites for the microcommunity fusion. The proposed algorithm is more efficient and has higher solution quality compared with other similar algorithms through the analysis of test results based on network data set.
Impulsive Neural Networks Algorithm Based on the Artificial Genome Model
Yuan Gao
2014-05-01
Full Text Available To describe gene regulatory networks, this article takes the framework of the artificial genome model and proposes impulsive neural networks algorithm based on the artificial genome model. Firstly, the gene expression and the cell division tree are applied to generate spiking neurons with specific attributes, neural network structure, connection weights and specific learning rules of each neuron. Next, the gene segment duplications and divergence model are applied to design the evolutionary algorithm of impulsive neural networks at the level of the artificial genome. The dynamic changes of developmental gene regulatory networks are controlled during the whole evolutionary process. Finally, the behavior of collecting food for autonomous intelligent agent is simulated, which is driven by nerves. Experimental results demonstrate that the algorithm in this article has the evolutionary ability on large-scale impulsive neural networks
Crime Busting Model Based on Dynamic Ranking Algorithms
Yang Cao
2013-01-01
Full Text Available This paper proposed a crime busting model with two dynamic ranking algorithms to detect the likelihood of a suspect and the possibility of a leader in a complex social network. Signally, in order to obtain the priority list of suspects, an advanced network mining approach with a dynamic cumulative nominating algorithm is adopted to rapidly reduce computational expensiveness than most other topology-based approaches. Our method can also greatly increase the accuracy of solution with the enhancement of semantic learning filtering at the same time. Moreover, another dynamic algorithm of node contraction is also presented to help identify the leader among conspirators. Test results are given to verify the theoretical results, which show the great performance for either small or large datasets.
An improved EZBC algorithm based on block bit length
Wang, Renlong; Ruan, Shuangchen; Liu, Chengxiang; Wang, Wenda; Zhang, Li
2011-12-01
Embedded ZeroBlock Coding and context modeling (EZBC) algorithm has high compression performance. However, it consumes large amounts of memory space because an Amplitude Quadtree of wavelet coefficients and other two link lists would be built during the encoding process. This is one of the big challenges for EZBC to be used in real time or hardware applications. An improved EZBC algorithm based on bit length of coefficients was brought forward in this article. It uses Bit Length Quadtree to complete the coding process and output the context for Arithmetic Coder. It can achieve the same compression performance as EZBC and save more than 75% memory space required in the encoding process. As Bit Length Quadtree can quickly locate the wavelet coefficients and judge their significance, the improved algorithm can dramatically accelerate the encoding speed. These improvements are also beneficial for hardware. PACS: 42.30.Va, 42.30.Wb
Pitch Based Wind Turbine Intelligent Speed Setpoint Adjustment Algorithms
Asier González-González
2014-06-01
Full Text Available This work is aimed at optimizing the wind turbine rotor speed setpoint algorithm. Several intelligent adjustment strategies have been investigated in order to improve a reward function that takes into account the power captured from the wind and the turbine speed error. After different approaches including Reinforcement Learning, the best results were obtained using a Particle Swarm Optimization (PSO-based wind turbine speed setpoint algorithm. A reward improvement of up to 10.67% has been achieved using PSO compared to a constant approach and 0.48% compared to a conventional approach. We conclude that the pitch angle is the most adequate input variable for the turbine speed setpoint algorithm compared to others such as rotor speed, or rotor angular acceleration.
CS-based fast ultrasound imaging with improved FISTA algorithm
Lin, Jie; He, Yugao; Shi, Guangming; Han, Tingyu
2015-08-01
In ultrasound imaging system, the wave emission and data acquisition is time consuming, which can be solved by adopting the plane wave as the transmitted signal, and the compressed sensing (CS) theory for data acquisition and image reconstruction. To overcome the very high computation complexity caused by introducing CS into ultrasound imaging, in this paper, we propose an improvement of the fast iterative shrinkage-thresholding algorithm (FISTA) to achieve the fast reconstruction of the ultrasound imaging, in which a modified setting is done with the parameter of step size for each iteration. Further, the GPU strategy is designed for the proposed algorithm, to guarantee the real time implementation of imaging. The simulation results show that the GPU-based image reconstruction algorithm can achieve the fast ultrasound imaging without damaging the quality of image.
A dynamic fuzzy clustering method based on genetic algorithm
ZHENG Yan; ZHOU Chunguang; LIANG Yanchun; GUO Dongwei
2003-01-01
A dynamic fuzzy clustering method is presented based on the genetic algorithm. By calculating the fuzzy dissimilarity between samples the essential associations among samples are modeled factually. The fuzzy dissimilarity between two samples is mapped into their Euclidean distance, that is, the high dimensional samples are mapped into the two-dimensional plane. The mapping is optimized globally by the genetic algorithm, which adjusts the coordinates of each sample, and thus the Euclidean distance, to approximate to the fuzzy dissimilarity between samples gradually. A key advantage of the proposed method is that the clustering is independent of the space distribution of input samples, which improves the flexibility and visualization. This method possesses characteristics of a faster convergence rate and more exact clustering than some typical clustering algorithms. Simulated experiments show the feasibility and availability of the proposed method.
A Reversible Image Steganographic Algorithm Based on Slantlet Transform
Sushil Kumar
2013-07-01
Full Text Available In this paper we present a reversible imagesteganography technique based on Slantlet transform (SLTand using advanced encryption standard (AES method. Theproposed method first encodes the message using two sourcecodes, viz., Huffman codes and a self-synchronizing variablelength code known as, T-code. Next, the encoded binarystring is encrypted using an improved AES method. Theencrypted data so obtained is embedded in the middle andhigh frequency sub-bands, obtained by applying 2-level ofSLT to the cover-image, using thresholding method. Theproposed algorithm is compared with the existing techniquesbased on wavelet transform. The Experimental results showthat the proposed algorithm can extract hidden message andrecover the original cover image with low distortion. Theproposed algorithm offers acceptable imperceptibility,security (two-layer security and provides robustness againstGaussian and Salt-n-Pepper noise attack.
Ant Colony Based Path Planning Algorithm for Autonomous Robotic Vehicles
Yogita Gigras
2012-11-01
Full Text Available The requirement of an autonomous robotic vehicles demand highly efficient algorithm as well as software. Today’s advanced computer hardware technology does not provide these types of extensive processing capabilities, so there is still a major space and time limitation for the technologies that are available for autonomous robotic applications. Now days, small to miniature mobile robots are required for investigation, surveillance and hazardous material detection for military and industrial applications. But these small sized robots have limited power capacity as well as memory and processing resources. A number of algorithms exist for producing optimal path for dynamically cost. This paper presents a new ant colony based approach which is helpful in solving path planning problem for autonomous robotic application. The experiment of simulation verified its validity of algorithm in terms of time.
An algorithm for motif-based network design
Mäki-Marttunen, Tuomo
2016-01-01
A determinant property of the structure of a biological network is the distribution of local connectivity patterns, i.e., network motifs. In this work, a method for creating directed, unweighted networks while promoting a certain combination of motifs is presented. This motif-based network algorithm starts with an empty graph and randomly connects the nodes by advancing or discouraging the formation of chosen motifs. The in- or out-degree distribution of the generated networks can be explicitly chosen. The algorithm is shown to perform well in producing networks with high occurrences of the targeted motifs, both ones consisting of 3 nodes as well as ones consisting of 4 nodes. Moreover, the algorithm can also be tuned to bring about global network characteristics found in many natural networks, such as small-worldness and modularity.
A Dither Modulation Audio Watermarking Algorithm Based on HAS
Yi-bo Huang
2012-11-01
Full Text Available In this study, we propose a dither modulation audio watermarking algorithm based on human auditory system which applied the theory of dither modulation. The algorithm made the two-value image watermarking to one-dimensional digital sequence firstly and used the Fibonacci to transform one-dimensional digital sequence. Then divide the audio into audio data segment and made discrete wavelet transform with audio data segment, every segment can adaptive choose quantization step. Finally put low frequency coefficients transformed embedding the watermarking which applied the dither modulation. When extract the watermark with no original audio, they realized blind extraction. The experimental results show that this algorithm has preferable robustness to against the attack from noise addition, compression, low pass filtering and re-sampling.
Eigenvalue based Spectrum Sensing Algorithms for Cognitive Radio
Zeng, Yonghong
2008-01-01
Spectrum sensing is a fundamental component is a cognitive radio. In this paper, we propose new sensing methods based on the eigenvalues of the covariance matrix of signals received at the secondary users. In particular, two sensing algorithms are suggested, one is based on the ratio of the maximum eigenvalue to minimum eigenvalue; the other is based on the ratio of the average eigenvalue to minimum eigenvalue. Using some latest random matrix theories (RMT), we quantify the distributions of these ratios and derive the probabilities of false alarm and probabilities of detection for the proposed algorithms. We also find the thresholds of the methods for a given probability of false alarm. The proposed methods overcome the noise uncertainty problem, and can even perform better than the ideal energy detection when the signals to be detected are highly correlated. The methods can be used for various signal detection applications without requiring the knowledge of signal, channel and noise power. Simulations based ...
Quantum-based algorithm for optimizing artificial neural networks.
Tzyy-Chyang Lu; Gwo-Ruey Yu; Jyh-Ching Juang
2013-08-01
This paper presents a quantum-based algorithm for evolving artificial neural networks (ANNs). The aim is to design an ANN with few connections and high classification performance by simultaneously optimizing the network structure and the connection weights. Unlike most previous studies, the proposed algorithm uses quantum bit representation to codify the network. As a result, the connectivity bits do not indicate the actual links but the probability of the existence of the connections, thus alleviating mapping problems and reducing the risk of throwing away a potential candidate. In addition, in the proposed model, each weight space is decomposed into subspaces in terms of quantum bits. Thus, the algorithm performs a region by region exploration, and evolves gradually to find promising subspaces for further exploitation. This is helpful to provide a set of appropriate weights when evolving the network structure and to alleviate the noisy fitness evaluation problem. The proposed model is tested on four benchmark problems, namely breast cancer and iris, heart, and diabetes problems. The experimental results show that the proposed algorithm can produce compact ANN structures with good generalization ability compared to other algorithms.
Localized Ambient Solidity Separation Algorithm Based Computer User Segmentation
Xiao Sun
2015-01-01
Full Text Available Most of popular clustering methods typically have some strong assumptions of the dataset. For example, the k-means implicitly assumes that all clusters come from spherical Gaussian distributions which have different means but the same covariance. However, when dealing with datasets that have diverse distribution shapes or high dimensionality, these assumptions might not be valid anymore. In order to overcome this weakness, we proposed a new clustering algorithm named localized ambient solidity separation (LASS algorithm, using a new isolation criterion called centroid distance. Compared with other density based isolation criteria, our proposed centroid distance isolation criterion addresses the problem caused by high dimensionality and varying density. The experiment on a designed two-dimensional benchmark dataset shows that our proposed LASS algorithm not only inherits the advantage of the original dissimilarity increments clustering method to separate naturally isolated clusters but also can identify the clusters which are adjacent, overlapping, and under background noise. Finally, we compared our LASS algorithm with the dissimilarity increments clustering method on a massive computer user dataset with over two million records that contains demographic and behaviors information. The results show that LASS algorithm works extremely well on this computer user dataset and can gain more knowledge from it.
Localized Ambient Solidity Separation Algorithm Based Computer User Segmentation.
Sun, Xiao; Zhang, Tongda; Chai, Yueting; Liu, Yi
2015-01-01
Most of popular clustering methods typically have some strong assumptions of the dataset. For example, the k-means implicitly assumes that all clusters come from spherical Gaussian distributions which have different means but the same covariance. However, when dealing with datasets that have diverse distribution shapes or high dimensionality, these assumptions might not be valid anymore. In order to overcome this weakness, we proposed a new clustering algorithm named localized ambient solidity separation (LASS) algorithm, using a new isolation criterion called centroid distance. Compared with other density based isolation criteria, our proposed centroid distance isolation criterion addresses the problem caused by high dimensionality and varying density. The experiment on a designed two-dimensional benchmark dataset shows that our proposed LASS algorithm not only inherits the advantage of the original dissimilarity increments clustering method to separate naturally isolated clusters but also can identify the clusters which are adjacent, overlapping, and under background noise. Finally, we compared our LASS algorithm with the dissimilarity increments clustering method on a massive computer user dataset with over two million records that contains demographic and behaviors information. The results show that LASS algorithm works extremely well on this computer user dataset and can gain more knowledge from it.
A Collaborative Neighbor Representation Based Face Recognition Algorithm
Zhengming Li
2013-01-01
Full Text Available We propose a new collaborative neighbor representation algorithm for face recognition based on a revised regularized reconstruction error (RRRE, called the two-phase collaborative neighbor representation algorithm (TCNR. Specifically, the RRRE is the division of l2-norm of reconstruction error of each class into a linear combination of l2-norm of reconstruction coefficients of each class, which can be used to increase the discrimination information for classification. The algorithm is as follows: in the first phase, the test sample is represented as a linear combination of all the training samples by incorporating the neighbor information into the objective function. In the second phase, we use the k classes to represent the test sample and calculate the collaborative neighbor representation coefficients. TCNR not only can preserve locality and similarity information of sparse coding but also can eliminate the side effect on the classification decision of the class that is far from the test sample. Moreover, the rationale and alternative scheme of TCNR are given. The experimental results show that TCNR algorithm achieves better performance than seven previous algorithms.
Incident Light Frequency-Based Image Defogging Algorithm
Wenbo Zhang
2017-01-01
Full Text Available To solve the color distortion problem produced by the dark channel prior algorithm, an improved method for calculating transmittance of all channels, respectively, was proposed in this paper. Based on the Beer-Lambert Law, the influence between the frequency of the incident light and the transmittance was analyzed, and the ratios between each channel’s transmittance were derived. Then, in order to increase efficiency, the input image was resized to a smaller size before acquiring the refined transmittance which will be resized to the same size of original image. Finally, all the transmittances were obtained with the help of the proportion between each color channel, and then they were used to restore the defogging image. Experiments suggest that the improved algorithm can produce a much more natural result image in comparison with original algorithm, which means the problem of high color saturation was eliminated. What is more, the improved algorithm speeds up by four to nine times compared to the original algorithm.
Object tracking algorithm based on contextual visual saliency
Fu, Bao; Peng, XianRong
2016-09-01
As to object tracking, the local context surrounding of the target could provide much effective information for getting a robust tracker. The spatial-temporal context (STC) learning algorithm proposed recently considers the information of the dense context around the target and has achieved a better performance. However STC only used image intensity as the object appearance model. But this appearance model not enough to deal with complicated tracking scenarios. In this paper, we propose a novel object appearance model learning algorithm. Our approach formulates the spatial-temporal relationships between the object of interest and its local context based on a Bayesian framework, which models the statistical correlation between high-level features (Circular-Multi-Block Local Binary Pattern) from the target and its surrounding regions. The tracking problem is posed by computing a visual saliency map, and obtaining the best target location by maximizing an object location likelihood function. Extensive experimental results on public benchmark databases show that our algorithm outperforms the original STC algorithm and other state-of-the-art tracking algorithms.
FACT. New image parameters based on the watershed-algorithm
Linhoff, Lena; Bruegge, Kai Arno; Buss, Jens [TU Dortmund (Germany). Experimentelle Physik 5b; Collaboration: FACT-Collaboration
2016-07-01
FACT, the First G-APD Cherenkov Telescope, is the first imaging atmospheric Cherenkov telescope that is using Geiger-mode avalanche photodiodes (G-APDs) as photo sensors. The raw data produced by this telescope are processed in an analysis chain, which leads to a classification of the primary particle that induce a shower and to an estimation of its energy. One important step in this analysis chain is the parameter extraction from shower images. By the application of a watershed algorithm to the camera image, new parameters are computed. Perceiving the brightness of a pixel as height, a set of pixels can be seen as 'landscape' with hills and valleys. A watershed algorithm groups all pixels to a cluster that belongs to the same hill. From the emerging segmented image, one can find new parameters for later analysis steps, e.g. number of clusters, their shape and containing photon charge. For FACT data, the FellWalker algorithm was chosen from the class of watershed algorithms, because it was designed to work on discrete distributions, in this case the pixels of a camera image. The FellWalker algorithm is implemented in FACT-tools, which provides the low level analysis framework for FACT. This talk will focus on the computation of new, FellWalker based, image parameters, which can be used for the gamma-hadron separation. Additionally, their distributions concerning real and Monte Carlo Data are compared.
Localized Ambient Solidity Separation Algorithm Based Computer User Segmentation
Sun, Xiao; Zhang, Tongda; Chai, Yueting; Liu, Yi
2015-01-01
Most of popular clustering methods typically have some strong assumptions of the dataset. For example, the k-means implicitly assumes that all clusters come from spherical Gaussian distributions which have different means but the same covariance. However, when dealing with datasets that have diverse distribution shapes or high dimensionality, these assumptions might not be valid anymore. In order to overcome this weakness, we proposed a new clustering algorithm named localized ambient solidity separation (LASS) algorithm, using a new isolation criterion called centroid distance. Compared with other density based isolation criteria, our proposed centroid distance isolation criterion addresses the problem caused by high dimensionality and varying density. The experiment on a designed two-dimensional benchmark dataset shows that our proposed LASS algorithm not only inherits the advantage of the original dissimilarity increments clustering method to separate naturally isolated clusters but also can identify the clusters which are adjacent, overlapping, and under background noise. Finally, we compared our LASS algorithm with the dissimilarity increments clustering method on a massive computer user dataset with over two million records that contains demographic and behaviors information. The results show that LASS algorithm works extremely well on this computer user dataset and can gain more knowledge from it. PMID:26221133
Remote triage support algorithm based on fuzzy logic.
Achkoski, Jugoslav; Koceski, S; Bogatinov, D; Temelkovski, B; Stevanovski, G; Kocev, I
2017-06-01
This paper presents a remote triage support algorithm as a part of a complex military telemedicine system which provides continuous monitoring of soldiers' vital sign data gathered on-site using unobtrusive set of sensors. The proposed fuzzy logic-based algorithm takes physiological data and classifies the casualties according to their health risk level, calculated following the Modified Early Warning Score (MEWS) methodology. To verify the algorithm, eight different evaluation scenarios using random vital sign data have been created. In each scenario, the hypothetical condition of the victims was assessed in parallel both by the system as well as by 50 doctors with significant experience in the field. The results showed that there is high (0.928) average correlation of the classification results. This suggests that the proposed algorithm can be used for automated remote triage in real life-saving situations even before the medical team arrives at the spot, and shorten the response times. Moreover, an additional study has been conducted in order to increase the computational efficiency of the algorithm, without compromising the quality of the classification results. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/.
Wideband Array Signal Detection Algorithm Based on Power Focusing
Gong Bin
2012-09-01
Full Text Available Aiming at the requirement of real-time signal detection in the passive surveillance system, a wideband array signal detection algorithm is proposed based on the concept of power focusing. By making use of the phase difference of the signal received by a uniform linear array, the algorithm makes the power of the received signal focused in the Direction Of Arrival (DOA with improved cascade FFT. Subsequently, the probability density function of the output noise at each angle is derived. Furthermore, a Constant False Alarm Rate (CFAR test statistic and the corresponding detection threshold are constructed. The theoretical probability of detection is also derived for different false alarm rate and Signal-to-Noise Ratio (SNR. The proposed algorithm is computationally efficient, and the detection process is independent of the prior information. Meanwhile, the results can act as the initial value for other algorithms with higher precision. Simulation results show that the proposed algorithm achieves good performance for weak signal detection.
A Greedy Algorithm for Neighborhood Overlap-Based Community Detection
Natarajan Meghanathan
2016-01-01
Full Text Available The neighborhood overlap (NOVER of an edge u-v is defined as the ratio of the number of nodes who are neighbors for both u and v to that of the number of nodes who are neighbors of at least u or v. In this paper, we hypothesize that an edge u-v with a lower NOVER score bridges two or more sets of vertices, with very few edges (other than u-v connecting vertices from one set to another set. Accordingly, we propose a greedy algorithm of iteratively removing the edges of a network in the increasing order of their neighborhood overlap and calculating the modularity score of the resulting network component(s after the removal of each edge. The network component(s that have the largest cumulative modularity score are identified as the different communities of the network. We evaluate the performance of the proposed NOVER-based community detection algorithm on nine real-world network graphs and compare the performance against the multi-level aggregation-based Louvain algorithm, as well as the original and time-efficient versions of the edge betweenness-based Girvan-Newman (GN community detection algorithm.
User Equilibrium Exchange Allocation Algorithm Based on Super Network
Peiyi Dong
2013-12-01
Full Text Available The theory of super network is an effective method to various traffic networks with means of multiple decision-making. It provides us with a favorable pricing decision tool for it combines a practical transport network with the space pricing decision. Spatial price equilibrium problem has always been the important research direction of the Transport Economics and regional transportation planning. As to how to combine the two, this paper presents the user equilibrium exchange allocation algorithm based on super network, which successfully keep the classical spatial price equilibrium problems (SPE into a super-network analysis framework. Through super-network analysis, we can add two virtual nodes in the network, which correspond to the virtual supply node and the super-super-demand virtual node, analysis the user equivalence with the SPE equilibrium and find the concrete steps of users exchange allocation algorithm based on super-network equilibrium. Finally, we carried out experiments to verify. The experiments show that: through the user equilibrium exchange SPE allocation algorithm based on super-network, we can get the steady-state equilibrium solution, which demonstrate that the algorithm is reasonable.
A CUDA-based reverse gridding algorithm for MR reconstruction.
Yang, Jingzhu; Feng, Chaolu; Zhao, Dazhe
2013-02-01
MR raw data collected using non-Cartesian method can be transformed on Cartesian grids by traditional gridding algorithm (GA) and reconstructed by Fourier transform. However, its runtime complexity is O(K×N(2)), where resolution of raw data is N×N and size of convolution window (CW) is K. And it involves a large number of matrix calculation including modulus, addition, multiplication and convolution. Therefore, a Compute Unified Device Architecture (CUDA)-based algorithm is proposed to improve the reconstruction efficiency of PROPELLER (a globally recognized non-Cartesian sampling method). Experiment shows a write-write conflict among multiple CUDA threads. This induces an inconsistent result when synchronously convoluting multiple k-space data onto the same grid. To overcome this problem, a reverse gridding algorithm (RGA) was developed. Different from the method of generating a grid window for each trajectory as in traditional GA, RGA calculates a trajectory window for each grid. This is what "reverse" means. For each k-space point in the CW, contribution is cumulated to this grid. Although this algorithm can be easily extended to reconstruct other non-Cartesian sampled raw data, we only implement it based on PROPELLER. Experiment illustrates that this CUDA-based RGA has successfully solved the write-write conflict and its reconstruction speed is 7.5 times higher than that of traditional GA.
Local Competition-Based Superpixel Segmentation Algorithm in Remote Sensing.
Liu, Jiayin; Tang, Zhenmin; Cui, Ying; Wu, Guoxing
2017-06-12
Remote sensing technologies have been widely applied in urban environments' monitoring, synthesis and modeling. Incorporating spatial information in perceptually coherent regions, superpixel-based approaches can effectively eliminate the "salt and pepper" phenomenon which is common in pixel-wise approaches. Compared with fixed-size windows, superpixels have adaptive sizes and shapes for different spatial structures. Moreover, superpixel-based algorithms can significantly improve computational efficiency owing to the greatly reduced number of image primitives. Hence, the superpixel algorithm, as a preprocessing technique, is more and more popularly used in remote sensing and many other fields. In this paper, we propose a superpixel segmentation algorithm called Superpixel Segmentation with Local Competition (SSLC), which utilizes a local competition mechanism to construct energy terms and label pixels. The local competition mechanism leads to energy terms locality and relativity, and thus, the proposed algorithm is less sensitive to the diversity of image content and scene layout. Consequently, SSLC could achieve consistent performance in different image regions. In addition, the Probability Density Function (PDF), which is estimated by Kernel Density Estimation (KDE) with the Gaussian kernel, is introduced to describe the color distribution of superpixels as a more sophisticated and accurate measure. To reduce computational complexity, a boundary optimization framework is introduced to only handle boundary pixels instead of the whole image. We conduct experiments to benchmark the proposed algorithm with the other state-of-the-art ones on the Berkeley Segmentation Dataset (BSD) and remote sensing images. Results demonstrate that the SSLC algorithm yields the best overall performance, while the computation time-efficiency is still competitive.
Landmark-Based Drift Compensation Algorithm for Inertial Pedestrian Navigation
Munoz Diaz, Estefania; Caamano, Maria; Fuentes Sánchez, Francisco Javier
2017-01-01
The navigation of pedestrians based on inertial sensors, i.e., accelerometers and gyroscopes, has experienced a great growth over the last years. However, the noise of medium- and low-cost sensors causes a high error in the orientation estimation, particularly in the yaw angle. This error, called drift, is due to the bias of the z-axis gyroscope and other slow changing errors, such as temperature variations. We propose a seamless landmark-based drift compensation algorithm that only uses inertial measurements. The proposed algorithm adds a great value to the state of the art, because the vast majority of the drift elimination algorithms apply corrections to the estimated position, but not to the yaw angle estimation. Instead, the presented algorithm computes the drift value and uses it to prevent yaw errors and therefore position errors. In order to achieve this goal, a detector of landmarks, i.e., corners and stairs, and an association algorithm have been developed. The results of the experiments show that it is possible to reliably detect corners and stairs using only inertial measurements eliminating the need that the user takes any action, e.g., pressing a button. Associations between re-visited landmarks are successfully made taking into account the uncertainty of the position. After that, the drift is computed out of all associations and used during a post-processing stage to obtain a low-drifted yaw angle estimation, that leads to successfully drift compensated trajectories. The proposed algorithm has been tested with quasi-error-free turn rate measurements introducing known biases and with medium-cost gyroscopes in 3D indoor and outdoor scenarios. PMID:28671622
Landmark-Based Drift Compensation Algorithm for Inertial Pedestrian Navigation.
Diaz, Estefania Munoz; Caamano, Maria; Sánchez, Francisco Javier Fuentes
2017-07-03
The navigation of pedestrians based on inertial sensors, i.e., accelerometers and gyroscopes, has experienced a great growth over the last years. However, the noise of medium- and low-cost sensors causes a high error in the orientation estimation, particularly in the yaw angle. This error, called drift, is due to the bias of the z-axis gyroscope and other slow changing errors, such as temperature variations. We propose a seamless landmark-based drift compensation algorithm that only uses inertial measurements. The proposed algorithm adds a great value to the state of the art, because the vast majority of the drift elimination algorithms apply corrections to the estimated position, but not to the yaw angle estimation. Instead, the presented algorithm computes the drift value and uses it to prevent yaw errors and therefore position errors. In order to achieve this goal, a detector of landmarks, i.e., corners and stairs, and an association algorithm have been developed. The results of the experiments show that it is possible to reliably detect corners and stairs using only inertial measurements eliminating the need that the user takes any action, e.g., pressing a button. Associations between re-visited landmarks are successfully made taking into account the uncertainty of the position. After that, the drift is computed out of all associations and used during a post-processing stage to obtain a low-drifted yaw angle estimation, that leads to successfully drift compensated trajectories. The proposed algorithm has been tested with quasi-error-free turn rate measurements introducing known biases and with medium-cost gyroscopes in 3D indoor and outdoor scenarios.
Effective pathfinding for four-wheeled robot based on combining Theta* and hybrid A* algorithms
Віталій Геннадійович Михалько
2016-07-01
Full Text Available Effective pathfinding algorithm based on Theta* and Hybrid A* algorithms was developed for four-wheeled robot. Pseudocode for algorithm was showed and explained. Algorithm and simulator for four-wheeled robot were implemented using Java programming language. Algorithm was tested on U-obstacles, complex maps and for parking problem
A Multiple-Neighborhood-Based Parallel Composite Local Search Algorithm for Timetable Problem
颜鹤; 郁松年
2004-01-01
This paper presents a parallel composite local search algorithm based on multiple search neighborhoods to solve a special kind of timetable problem. The new algorithm can also effectively solve those problems that can be solved by general local search algorithms. Experimental results show that the new algorithm can generate better solutions than general local search algorithms.
New Iterative Learning Control Algorithms Based on Vector Plots Analysis1）
XIESheng-Li; TIANSen-Ping; XIEZhen-Dong
2004-01-01
Based on vector plots analysis, this paper researches the geometric frame of iterativelearning control method. New structure of iterative learning algorithms is obtained by analyzingthe vector plots of some general algorithms. The structure of the new algorithm is different fromthose of the present algorithms. It is of faster convergence speed and higher accuracy. Simulationspresented here illustrate the effectiveness and advantage of the new algorithm.
Measurement Theory in Deutsch's Algorithm Based on the Truth Values
Nagata, Koji; Nakamura, Tadao
2016-08-01
We propose a new measurement theory, in qubits handling, based on the truth values, i.e., the truth T (1) for true and the falsity F (0) for false. The results of measurement are either 0 or 1. To implement Deutsch's algorithm, we need both observability and controllability of a quantum state. The new measurement theory can satisfy these two. Especially, we systematically describe our assertion based on more mathematical analysis using raw data in a thoughtful experiment.
Immune and Genetic Algorithm Based Assembly Sequence Planning
YANG Jian-guo; LI Bei-zhi; YU Lei; JIN Yu-song
2004-01-01
In this paper an assembly sequence planning model inspired by natural immune and genetic algorithm (ASPIG) based on the part degrees of freedom matrix (PDFM) is proposed, and a proto system - DSFAS based on the ASPIG is introduced to solve assembly sequence problem. The concept and generation of PDFM and DSFAS are also discussed. DSFAS can prevent premature convergence, and promote population diversity, and can accelerate the learning and convergence speed in behavior evolution problem.
TOA-BASED ROBUST LOCATION ALGORITHMS FOR WIRELESS CELLULAR NETWORKS
Sun Guolin; Guo Wei
2005-01-01
Caused by Non-Line-Of-Sight (NLOS) propagation effect, the non-symmetric contamination of measured Time Of Arrival (TOA) data leads to high inaccuracies of the conventional TOA based mobile location techniques. Robust position estimation method based on bootstrapping M-estimation and Huber estimator are proposed to mitigate the effects of NLOS propagation on the location error. Simulation results show the improvement over traditional Least-Square (LS)algorithm on location accuracy under different channel environments.
75 FR 44053 - Proposed Collection; Comment Request: CDFI/CDE Project Profiles Web Form
2010-07-27
... description may be for a project previously reported to the CDFI Fund through the Community Investment Impact... Community Development Financial Institutions Fund Proposed Collection; Comment Request: CDFI/CDE Project... concerning the CDFI/CDE Project Profile Web Form, a voluntary information collection effort involving...
A Novel Assembly Line Balancing Method Based on PSO Algorithm
Xiaomei Hu
2014-01-01
Full Text Available Assembly line is widely used in manufacturing system. Assembly line balancing problem is a crucial question during design and management of assembly lines since it directly affects the productivity of the whole manufacturing system. The model of assembly line balancing problem is put forward and a general optimization method is proposed. The key data on assembly line balancing problem is confirmed, and the precedence relations diagram is described. A double objective optimization model based on takt time and smoothness index is built, and balance optimization scheme based on PSO algorithm is proposed. Through the simulation experiments of examples, the feasibility and validity of the assembly line balancing method based on PSO algorithm is proved.
A self region based real-valued negative selection algorithm
ZHANG Feng-bin; WANG Da-wei; WANG Sheng-wen
2008-01-01
Point-wise negative selection algorithms, which generate their detector sets based on point of self da-ta, have lower training efficiency and detection rate. To solve this problem, a self region based real-valued neg-ative selection algorithm is presented. In this new approach, the continuous self region is defined by the collec-tion of self data, the partial training takes place at the training stage according to both the radius of self region and the cosine distance between gravity of the self region and detector candidate, and variable detectors in the self region are deployed. The algorithm is tested using the triangle shape of self region in the 2-D complement space and KDD CUP 1999 data set. Results show that, more information can be provided when the training self points are used together as a whole, and compared with the point-wise negative selection algorithm, the new ap-proach can improve the training efficiency of system and the detection rate significantly.
Entropy-Based Search Algorithm for Experimental Design
Malakar, N. K.; Knuth, K. H.
2011-03-01
The scientific method relies on the iterated processes of inference and inquiry. The inference phase consists of selecting the most probable models based on the available data; whereas the inquiry phase consists of using what is known about the models to select the most relevant experiment. Optimizing inquiry involves searching the parameterized space of experiments to select the experiment that promises, on average, to be maximally informative. In the case where it is important to learn about each of the model parameters, the relevance of an experiment is quantified by Shannon entropy of the distribution of experimental outcomes predicted by a probable set of models. If the set of potential experiments is described by many parameters, we must search this high-dimensional entropy space. Brute force search methods will be slow and computationally expensive. We present an entropy-based search algorithm, called nested entropy sampling, to select the most informative experiment for efficient experimental design. This algorithm is inspired by Skilling's nested sampling algorithm used in inference and borrows the concept of a rising threshold while a set of experiment samples are maintained. We demonstrate that this algorithm not only selects highly relevant experiments, but also is more efficient than brute force search. Such entropic search techniques promise to greatly benefit autonomous experimental design.
Family genetic algorithms based on gene exchange and its application
Li Jianhua; Ding Xiangqian; Wang Sunan; Yu Qing
2006-01-01
Genetic Algorithms (GA) are a search techniques based on mechanics of nature selection and have already been successfully applied in many diverse areas. However, increasing samples show that GA's performance is not as good as it was expected to be. Criticism of this algorithm includes the slow speed and premature result during convergence procedure. In order to improve the performance, the population size and individuals' space is emphatically described. The influence of individuals' space and population size on the operators is analyzed. And a novel family genetic algorithm (FGA) is put forward based on this analysis. In this novel algorithm, the optimum solution families closed to quality individuals is constructed, which is exchanged found by a search in the world space. Search will be done in this microspace. The family that can search better genes in a limited period of time would win a new life. At the same time, the best gene of this micro space with the basic population in the world space is exchanged. Finally, the FGA is applied to the function optimization and image matching through several experiments. The results show that the FGA possessed high performance.
Development of antibiotic regimens using graph based evolutionary algorithms.
Corns, Steven M; Ashlock, Daniel A; Bryden, Kenneth M
2013-12-01
This paper examines the use of evolutionary algorithms in the development of antibiotic regimens given to production animals. A model is constructed that combines the lifespan of the animal and the bacteria living in the animal's gastro-intestinal tract from the early finishing stage until the animal reaches market weight. This model is used as the fitness evaluation for a set of graph based evolutionary algorithms to assess the impact of diversity control on the evolving antibiotic regimens. The graph based evolutionary algorithms have two objectives: to find an antibiotic treatment regimen that maintains the weight gain and health benefits of antibiotic use and to reduce the risk of spreading antibiotic resistant bacteria. This study examines different regimens of tylosin phosphate use on bacteria populations divided into Gram positive and Gram negative types, with a focus on Campylobacter spp. Treatment regimens were found that provided decreased antibiotic resistance relative to conventional methods while providing nearly the same benefits as conventional antibiotic regimes. By using a graph to control the information flow in the evolutionary algorithm, a variety of solutions along the Pareto front can be found automatically for this and other multi-objective problems.
aTrunk—An ALS-Based Trunk Detection Algorithm
Sebastian Lamprecht
2015-08-01
Full Text Available This paper presents a rapid multi-return ALS-based (Airborne Laser Scanning tree trunk detection approach. The multi-core Divide & Conquer algorithm uses a CBH (Crown Base Height estimation and 3D-clustering approach to isolate points associated with single trunks. For each trunk, a principal-component-based linear model is fitted, while a deterministic modification of LO-RANSAC is used to identify an optimal model. The algorithm returns a vector-based model for each identified trunk while parameters like the ground position, zenith orientation, azimuth orientation and length of the trunk are provided. The algorithm performed well for a study area of 109 trees (about 2/3 Norway Spruce and 1/3 European Beech, with a point density of 7.6 points per m2, while a detection rate of about 75% and an overall accuracy of 84% were reached. Compared to crown-based tree detection methods, the aTrunk approach has the advantages of a high reliability (5% commission error and its high tree positioning accuracy (0.59m average difference and 0.78m RMSE. The usage of overlapping segments with parametrizable size allows a seamless detection of the tree trunks.
A Survey Paper on Deduplication by Using Genetic Algorithm Alongwith Hash-Based Algorithm
Miss. J. R. Waykole
2014-01-01
Full Text Available In today‟s world, by increasing the volume of information available in digital libraries, most of the system may be affected by the existence of replicas in their warehouses. This is due to the fact that, clean and replica-free warehouse not only allow the retrieval of information which is of higher quality but also lead to more concise data and reduces computational time and resources to process this data. Here, we propose a genetic programming approach along with hash-based similarity i.e, with MD5 and SHA-1 algorithm. This approach removes the replicas data and finds the optimization solution to deduplication of records.
Staff line detection and revision algorithm based on subsection projection and correlation algorithm
Yang, Yin-xian; Yang, Ding-li
2013-03-01
Staff line detection plays a key role in OMR technology, and is the precon-ditions of subsequent segmentation 1& recognition of music sheets. For the phenomena of horizontal inclination & curvature of staff lines and vertical inclination of image, which often occur in music scores, an improved approach based on subsection projection is put forward to realize the detection of original staff lines and revision in an effect to implement staff line detection more successfully. Experimental results show the presented algorithm can detect and revise staff lines fast and effectively.
A test sheet generating algorithm based on intelligent genetic algorithm and hierarchical planning
Gu, Peipei; Niu, Zhendong; Chen, Xuting; Chen, Wei
2013-03-01
In recent years, computer-based testing has become an effective method to evaluate students' overall learning progress so that appropriate guiding strategies can be recommended. Research has been done to develop intelligent test assembling systems which can automatically generate test sheets based on given parameters of test items. A good multisubject test sheet depends on not only the quality of the test items but also the construction of the sheet. Effective and efficient construction of test sheets according to multiple subjects and criteria is a challenging problem. In this paper, a multi-subject test sheet generation problem is formulated and a test sheet generating approach based on intelligent genetic algorithm and hierarchical planning (GAHP) is proposed to tackle this problem. The proposed approach utilizes hierarchical planning to simplify the multi-subject testing problem and adopts genetic algorithm to process the layered criteria, enabling the construction of good test sheets according to multiple test item requirements. Experiments are conducted and the results show that the proposed approach is capable of effectively generating multi-subject test sheets that meet specified requirements and achieve good performance.
A new JPEG-based steganographic algorithm for mobile devices
Agaian, Sos S.; Cherukuri, Ravindranath C.; Schneider, Erik C.; White, Gregory B.
2006-05-01
Currently, cellular phones constitute a significant portion of the global telecommunications market. Modern cellular phones offer sophisticated features such as Internet access, on-board cameras, and expandable memory which provide these devices with excellent multimedia capabilities. Because of the high volume of cellular traffic, as well as the ability of these devices to transmit nearly all forms of data. The need for an increased level of security in wireless communications is becoming a growing concern. Steganography could provide a solution to this important problem. In this article, we present a new algorithm for JPEG-compressed images which is applicable to mobile platforms. This algorithm embeds sensitive information into quantized discrete cosine transform coefficients obtained from the cover JPEG. These coefficients are rearranged based on certain statistical properties and the inherent processing and memory constraints of mobile devices. Based on the energy variation and block characteristics of the cover image, the sensitive data is hidden by using a switching embedding technique proposed in this article. The proposed system offers high capacity while simultaneously withstanding visual and statistical attacks. Based on simulation results, the proposed method demonstrates an improved retention of first-order statistics when compared to existing JPEG-based steganographic algorithms, while maintaining a capacity which is comparable to F5 for certain cover images.
Variance-based fingerprint distance adjustment algorithm for indoor localization
Xiaolong Xu; Yu Tang; Xinheng Wang; Yun Zhang
2015-01-01
The multipath effect and movements of people in in-door environments lead to inaccurate localization. Through the test, calculation and analysis on the received signal strength in-dication (RSSI) and the variance of RSSI, we propose a novel variance-based fingerprint distance adjustment algorithm (VFDA). Based on the rule that variance decreases with the increase of RSSI mean, VFDA calculates RSSI variance with the mean value of received RSSIs. Then, we can get the correction weight. VFDA adjusts the fingerprint distances with the correction weight based on the variance of RSSI, which is used to correct the fingerprint distance. Besides, a threshold value is applied to VFDA to im-prove its performance further. VFDA and VFDA with the threshold value are applied in two kinds of real typical indoor environments deployed with several Wi-Fi access points. One is a quadrate lab room, and the other is a long and narrow corridor of a building. Experimental results and performance analysis show that in in-door environments, both VFDA and VFDA with the threshold have better positioning accuracy and environmental adaptability than the current typical positioning methods based on the k-nearest neighbor algorithm and the weighted k-nearest neighbor algorithm with similar computational costs.
Multi-objective community detection based on memetic algorithm.
Peng Wu
Full Text Available Community detection has drawn a lot of attention as it can provide invaluable help in understanding the function and visualizing the structure of networks. Since single objective optimization methods have intrinsic drawbacks to identifying multiple significant community structures, some methods formulate the community detection as multi-objective problems and adopt population-based evolutionary algorithms to obtain multiple community structures. Evolutionary algorithms have strong global search ability, but have difficulty in locating local optima efficiently. In this study, in order to identify multiple significant community structures more effectively, a multi-objective memetic algorithm for community detection is proposed by combining multi-objective evolutionary algorithm with a local search procedure. The local search procedure is designed by addressing three issues. Firstly, nondominated solutions generated by evolutionary operations and solutions in dominant population are set as initial individuals for local search procedure. Then, a new direction vector named as pseudonormal vector is proposed to integrate two objective functions together to form a fitness function. Finally, a network specific local search strategy based on label propagation rule is expanded to search the local optimal solutions efficiently. The extensive experiments on both artificial and real-world networks evaluate the proposed method from three aspects. Firstly, experiments on influence of local search procedure demonstrate that the local search procedure can speed up the convergence to better partitions and make the algorithm more stable. Secondly, comparisons with a set of classic community detection methods illustrate the proposed method can find single partitions effectively. Finally, the method is applied to identify hierarchical structures of networks which are beneficial for analyzing networks in multi-resolution levels.
Fuzzy Sets-based Control Rules for Terminating Algorithms
Jose L. VERDEGAY
2002-01-01
Full Text Available In this paper some problems arising in the interface between two different areas, Decision Support Systems and Fuzzy Sets and Systems, are considered. The Model-Base Management System of a Decision Support System which involves some fuzziness is considered, and in that context the questions on the management of the fuzziness in some optimisation models, and then of using fuzzy rules for terminating conventional algorithms are presented, discussed and analyzed. Finally, for the concrete case of the Travelling Salesman Problem, and as an illustration of determination, management and using the fuzzy rules, a new algorithm easy to implement in the Model-Base Management System of any oriented Decision Support System is shown.
Memoryless cooperative graph search based on the simulated annealing algorithm
Hou Jian; Yan Gang-Feng; Fan Zhen
2011-01-01
We have studied the problem of reaching a globally optimal segment for a graph-like environment with a single or a group of autonomous mobile agents. Firstly, two efficient simulated-annealing-like algorithms are given for a single agent to solve the problem in a partially known environment and an unknown environment, respectively. It shows that under both proposed control strategies, the agent will eventually converge to a globally optimal segment with probability 1Secondly, we use multi-agent searching to simultaneously reduce the computation complexity and accelerate convergence based on the algorithms we have given for a single agent. By exploiting graph partition, a gossip consensus method based scheme is presented to update the key parameter-radius of the graph, ensuring that the agents spend much less time finding a globally optimal segment.
Voronoi-based localisation algorithm for mobile sensor networks
Guan, Zixiao; Zhang, Yongtao; Zhang, Baihai; Dong, Lijing
2016-11-01
Localisation is an essential and important part in wireless sensor networks (WSNs). Many applications require location information. So far, there are less researchers studying on mobile sensor networks (MSNs) than static sensor networks (SSNs). However, MSNs are required in more and more areas such that the number of anchor nodes can be reduced and the location accuracy can be improved. In this paper, we firstly propose a range-free Voronoi-based Monte Carlo localisation algorithm (VMCL) for MSNs. We improve the localisation accuracy by making better use of the information that a sensor node gathers. Then, we propose an optimal region selection strategy of Voronoi diagram based on VMCL, called ORSS-VMCL, to increase the efficiency and accuracy for VMCL by adapting the size of Voronoi area during the filtering process. Simulation results show that the accuracy of these two algorithms, especially ORSS-VMCL, outperforms traditional MCL.
Genetic Algorithm Based Hybrid Fuzzy System for Assessing Morningness
Animesh Biswas
2014-01-01
Full Text Available This paper describes a real life case example on the assessment process of morningness of individuals using genetic algorithm based hybrid fuzzy system. It is observed that physical and mental performance of human beings in different time slots of a day are majorly influenced by morningness orientation of those individuals. To measure the morningness of people various self-reported questionnaires were developed by different researchers in the past. Among them reduced version of Morningness-Eveningness Questionnaire is mostly accepted. Almost all of the linguistic terms used in questionnaires are fuzzily defined. So, assessing them in crisp environments with their responses does not seem to be justifiable. Fuzzy approach based research works for assessing morningness of people are very few in the literature. In this paper, genetic algorithm is used to tune the parameters of a Mamdani fuzzy inference model to minimize error with their predicted outputs for assessing morningness of people.
Fuzzy Controllers Based Multipath Routing Algorithm in MANET
Pi, Shangchao; Sun, Baolin
Mobile ad hoc networks (MANETs) consist of a collection of wireless mobile nodes which dynamically exchange data among themselves without the reliance on a fixed base station or a wired backbone network. Due to the limited transmission range of wireless network nodes, multiple hops are usually needed for a node to exchange information with any other node in the network. Multipath routing allows the establishment of multiple paths between a single source and single destination node. The multipath routing in mobile ad hoc networks is difficult because the network topology may change constantly, and the available alternative path is inherently unreliable. This paper introduces a fuzzy controllers based multipath routing algorithm in MANET (FMRM). The key idea of FMRM algorithm is to construct the fuzzy controllers with the help to reduce reconstructions in the ad hoc network. The simulation results show that the proposed approach is effective and efficient in applications to the MANETs. It is an available approach to multipath routing decision.
A Syntactic Classification based Web Page Ranking Algorithm
Mukhopadhyay, Debajyoti; Kim, Young-Chon
2011-01-01
The existing search engines sometimes give unsatisfactory search result for lack of any categorization of search result. If there is some means to know the preference of user about the search result and rank pages according to that preference, the result will be more useful and accurate to the user. In the present paper a web page ranking algorithm is being proposed based on syntactic classification of web pages. Syntactic Classification does not bother about the meaning of the content of a web page. The proposed approach mainly consists of three steps: select some properties of web pages based on user's demand, measure them, and give different weightage to each property during ranking for different types of pages. The existence of syntactic classification is supported by running fuzzy c-means algorithm and neural network classification on a set of web pages. The change in ranking for difference in type of pages but for same query string is also being demonstrated.
Research on image self-recovery algorithm based on DCT
Shengbing Che
2010-06-01
Full Text Available Image compression operator based on discrete cosine transform was brought up. A securer scrambling locational operator was put forward based on the concept of anti-tamper radius. The basic idea of the algorithm is that it first combined image block compressed data with eigenvalue of image block and its offset block, then scrambled or encrypted and embeded them into least significant bit of corresponding offset block. This algorithm could pinpoint tampered image block and tampering type accurately. It could recover tampered block with good image quality when tamper occured within the limits of the anti-tamper radius. It could effectively resist vector quantization and synchronous counterfeiting attacks on self-embedding watermarking schemes.
Quantum Cryptography Based on the Deutsch-Jozsa Algorithm
Nagata, Koji; Nakamura, Tadao; Farouk, Ahmed
2017-09-01
Recently, secure quantum key distribution based on Deutsch's algorithm using the Bell state is reported (Nagata and Nakamura, Int. J. Theor. Phys. doi: 10.1007/s10773-017-3352-4, 2017). Our aim is of extending the result to a multipartite system. In this paper, we propose a highly speedy key distribution protocol. We present sequre quantum key distribution based on a special Deutsch-Jozsa algorithm using Greenberger-Horne-Zeilinger states. Bob has promised to use a function f which is of one of two kinds; either the value of f( x) is constant for all values of x, or else the value of f( x) is balanced, that is, equal to 1 for exactly half of the possible x, and 0 for the other half. Here, we introduce an additional condition to the function when it is balanced. Our quantum key distribution overcomes a classical counterpart by a factor O(2 N ).
The positioning algorithm based on feature variance of billet character
Yi, Jiansong; Hong, Hanyu; Shi, Yu; Chen, Hongyang
2015-12-01
In the process of steel billets recognition on the production line, the key problem is how to determine the position of the billet from complex scenes. To solve this problem, this paper presents a positioning algorithm based on the feature variance of billet character. Using the largest intra-cluster variance recursive method based on multilevel filtering, the billet characters are segmented completely from the complex scenes. There are three rows of characters on each steel billet, we are able to determine whether the connected regions, which satisfy the condition of the feature variance, are on a straight line. Then we can accurately locate the steel billet. The experimental results demonstrated that the proposed method in this paper is competitive to other methods in positioning the characters and it also reduce the running time. The algorithm can provide a better basis for the character recognition.
Manifold learning based registration algorithms applied to multimodal images.
Azampour, Mohammad Farid; Ghaffari, Aboozar; Hamidinekoo, Azam; Fatemizadeh, Emad
2014-01-01
Manifold learning algorithms are proposed to be used in image processing based on their ability in preserving data structures while reducing the dimension and the exposure of data structure in lower dimension. Multi-modal images have the same structure and can be registered together as monomodal images if only structural information is shown. As a result, manifold learning is able to transform multi-modal images to mono-modal ones and subsequently do the registration using mono-modal methods. Based on this application, in this paper novel similarity measures are proposed for multi-modal images in which Laplacian eigenmaps are employed as manifold learning algorithm and are tested against rigid registration of PET/MR images. Results show the feasibility of using manifold learning as a way of calculating the similarity between multimodal images.
Multicast Routing Problem Using Tree-Based Cuckoo Optimization Algorithm
Mahmood Sardarpour
2016-06-01
Full Text Available The problem of QoS multicast routing is to find a multicast tree with the least expense/cost which would meet the limitations such as band width, delay and loss rate. This is a NP-Complete problem. To solve the problem of multicast routing, the entire routes from the source node to every destination node are often recognized. Then the routes are integrated and changed into a single multicast tree. But they are slow and complicated methods. The present paper introduces a new tree-based optimization method to overcome such weaknesses. The recommended method directly optimizes the multicast tree. Therefore a tree-based typology including several spanning trees is created which combines the trees two by two. For this purpose, the Cuckoo Algorithm is used which is proved to be well converged and makes quick calculations. The simulation conducted on different types of network typologies proved that it is a practical and influential algorithm.
Genetic algorithm for network cost minimization using threshold based discounting
Hrvoje Podnar
2003-01-01
Full Text Available We present a genetic algorithm for heuristically solving a cost minimization problem applied to communication networks with threshold based discounting. The network model assumes that every two nodes can communicate and offers incentives to combine flow from different sources. Namely, there is a prescribed threshold on every link, and if the total flow on a link is greater than the threshold, the cost of this flow is discounted by a factor α. A heuristic algorithm based on genetic strategy is developed and applied to a benchmark set of problems. The results are compared with former branch and bound results using the CPLEX® solver. For larger data instances we were able to obtain improved solutions using less CPU time, confirming the effectiveness of our heuristic approach.
A New Evolutionary Algorithm Based on the Decimal Coding
无
2002-01-01
Traditional Evolutionary Algorithm (EAs) is based on the binary code, real number code, structure code and so on. But these coding strategies have their own advantages and disadvantages for the optimization of functions. In this paper a new Decimal Coding Strategy (DCS) ,which is convenient for space division and alterable precision, was proposed, and the theory analysis of its implicit parallelism and convergence was also discussed. We also redesign several genetic operators for the decimal code. In order to utilize the historical information of the existing individuals in the process of evolution and avoid repeated exploring,the strategies of space shrinking and precision alterable, are adopted. Finally, the evolutionary algorithm based on decimal coding (DCEAs) was applied to the optimization of functions, the optimization of parameter, mixed-integer nonlinear programming. Comparison with traditional GAs was made and the experimental results show that the performances of DCEAS are better than the tradition GAs.
Unified HMM-based layout analysis framework and algorithm
陈明; 丁晓青; 吴佑寿
2003-01-01
To manipulate the layout analysis problem for complex or irregular document image, a Unified HMM-based Layout Analysis Framework is presented in this paper. Based on the multi-resolution wavelet analysis results of the document image, we use HMM method in both inner-scale image model and trans-scale context model to classify the pixel region properties, such as text, picture or background. In each scale, a HMM direct segmentation method is used to get better inner-scale classification result. Then another HMM method is used to fuse the inner-scale result in each scale and then get better final seg- mentation result. The optimized algorithm uses a stop rule in the coarse to fine multi-scale segmentation process, so the speed is improved remarkably. Experiments prove the efficiency of proposed algorithm.
Mlambo, CS
2015-01-01
Full Text Available In this paper, implementations of three Hough Transform based fingerprint alignment algorithms are analyzed with respect to time complexity on Java Card environment. Three algorithms are: Local Match Based Approach (LMBA), Discretized Rotation Based...
Computer-Based Algorithmic Determination of Muscle Movement Onset Using M-Mode Ultrasonography
2017-04-01
from baseline. Computerized algorithms Computed MO was determined by three separate classes of algorithms using RStudio: (i) a novel standard...ARL-RP-0596 ● APR 2017 US Army Research Laboratory Computer -Based Algorithmic Determination of Muscle Movement Onset Using M...the originator. ARL-RP-0596 ● APR 2017 US Army Research Laboratory Computer -Based Algorithmic Determination of Muscle Movement
Variable Step Size Maximum Correntropy Criteria Based Adaptive Filtering Algorithm
S. Radhika
2016-04-01
Full Text Available Maximum correntropy criterion (MCC based adaptive filters are found to be robust against impulsive interference. This paper proposes a novel MCC based adaptive filter with variable step size in order to obtain improved performance in terms of both convergence rate and steady state error with robustness against impulsive interference. The optimal variable step size is obtained by minimizing the Mean Square Deviation (MSD error from one iteration to the other. Simulation results in the context of a highly impulsive system identification scenario show that the proposed algorithm has faster convergence and lesser steady state error than the conventional MCC based adaptive filters.
A novel pipeline based FPGA implementation of a genetic algorithm
Thirer, Nonel
2014-05-01
To solve problems when an analytical solution is not available, more and more bio-inspired computation techniques have been applied in the last years. Thus, an efficient algorithm is the Genetic Algorithm (GA), which imitates the biological evolution process, finding the solution by the mechanism of "natural selection", where the strong has higher chances to survive. A genetic algorithm is an iterative procedure which operates on a population of individuals called "chromosomes" or "possible solutions" (usually represented by a binary code). GA performs several processes with the population individuals to produce a new population, like in the biological evolution. To provide a high speed solution, pipelined based FPGA hardware implementations are used, with a nstages pipeline for a n-phases genetic algorithm. The FPGA pipeline implementations are constraints by the different execution time of each stage and by the FPGA chip resources. To minimize these difficulties, we propose a bio-inspired technique to modify the crossover step by using non identical twins. Thus two of the chosen chromosomes (parents) will build up two new chromosomes (children) not only one as in classical GA. We analyze the contribution of this method to reduce the execution time in the asynchronous and synchronous pipelines and also the possibility to a cheaper FPGA implementation, by using smaller populations. The full hardware architecture for a FPGA implementation to our target ALTERA development card is presented and analyzed.
Multiobjective Optimization Method Based on Adaptive Parameter Harmony Search Algorithm
P. Sabarinath
2015-01-01
Full Text Available The present trend in industries is to improve the techniques currently used in design and manufacture of products in order to meet the challenges of the competitive market. The crucial task nowadays is to find the optimal design and machining parameters so as to minimize the production costs. Design optimization involves more numbers of design variables with multiple and conflicting objectives, subjected to complex nonlinear constraints. The complexity of optimal design of machine elements creates the requirement for increasingly effective algorithms. Solving a nonlinear multiobjective optimization problem requires significant computing effort. From the literature it is evident that metaheuristic algorithms are performing better in dealing with multiobjective optimization. In this paper, we extend the recently developed parameter adaptive harmony search algorithm to solve multiobjective design optimization problems using the weighted sum approach. To determine the best weightage set for this analysis, a performance index based on least average error is used to determine the index of each weightage set. The proposed approach is applied to solve a biobjective design optimization of disc brake problem and a newly formulated biobjective design optimization of helical spring problem. The results reveal that the proposed approach is performing better than other algorithms.
Computer Crime Forensics Based on Improved Decision Tree Algorithm
Ying Wang
2014-04-01
Full Text Available To find out the evidence of crime-related evidence and association rules among massive data, the classic decision tree algorithms such as ID3 for classification analysis have appeared in related prototype systems. So how to make it more suitable for computer forensics in variable environments becomes a hot issue. When selecting classification attributes, ID3 relies on computation of information entropy. Then the attributes owning more value are selected as classification nodes of the decision tress. Such classification is unrealistic under many cases. During the process of ID3 algorithm there are too many logarithms, so it is complicated to handle with the dataset which has various classification attributes. Therefore, contraposing the special demand for computer crime forensics, ID3 algorithm is improved and a novel classification attribute selection method based on Maclaurin-Priority Value First method is proposed. It adopts the foot changing formula and infinitesimal substitution to simplify the logarithms in ID3. For the errors generated in this process, an apposite constant is introduced to be multiplied by the simplified formulas for compensation. The idea of Priority Value First is introduced to solve the problems of value deviation. The performance of improved method is strictly proved in theory. Finally, the experiments verify that our scheme has advantage in computation time and classification accuracy, compared to ID3 and two existing algorithms
Broadcast Networks based on the Virus Evolutionary Algorithm
Jian-xin Zhu
2014-05-01
Full Text Available An optimization algorithm for virus evolution is to research the spread process of a computer or biological virus in network system. The objective of the algorithm is mainly to control the speed of the virus evolution with limited network resource and to study how users can be infected in the network. A dynamical probabilistic system on a connected graph is adopted to model the virus evolution. A traditional virus evolution model needs to solve a non-convex optimization problem taking the spectral radius function of a nonnegative matrix as an optimization objective in the description of virus evolution model. On this basis, two novel approximation algorithms are proposed in this paper. Based on continuous convex approximation, the first one is a suboptimal with rapid speed. The second one can adopt branch-and-bound techniques to achieve a global optimal solution, which use some key inequalities of nonnegative matrix. Comparing with traditional virus evolution model, the simulation experiment shows that the improved algorithm can reach the global optimum in the process of virus evolution and has fast convergence capability in different network conditions.
Parameter Optimization of Linear Quadratic Controller Based on Genetic Algorithm
LI Jimin; SHANG Chaoxuan; ZOU Minghu
2007-01-01
The selection of weighting matrix in design of the linear quadratic optimal controller is an important topic in the control theory. In this paper, an approach based on genetic algorithm is presented for selecting the weighting matrix for the optimal controller. Genetic algorithm is adaptive heuristic search algorithm premised on the evolutionary ideas of natural selection and genetic. In this algorithm, the fitness function is used to evaluate individuals and reproductive success varies with fitness. In the design of the linear quadratic optimal controller, the fitness function has relation to the anticipated step response of the system. Not only can the controller designed by this approach meet the demand of the performance indexes of linear quadratic controller, but also satisfy the anticipated step response of close-loop system. The method possesses a higher calculating efficiency and provides technical support for the optimal controller in engineering application. The simulation of a three-order single-input single-output (SISO) system has demonstrated the feasibility and validity of the approach.
A haplotype inference algorithm for trios based on deterministic sampling
Iliadis Alexandros
2010-08-01
Full Text Available Abstract Background In genome-wide association studies, thousands of individuals are genotyped in hundreds of thousands of single nucleotide polymorphisms (SNPs. Statistical power can be increased when haplotypes, rather than three-valued genotypes, are used in analysis, so the problem of haplotype phase inference (phasing is particularly relevant. Several phasing algorithms have been developed for data from unrelated individuals, based on different models, some of which have been extended to father-mother-child "trio" data. Results We introduce a technique for phasing trio datasets using a tree-based deterministic sampling scheme. We have compared our method with publicly available algorithms PHASE v2.1, BEAGLE v3.0.2 and 2SNP v1.7 on datasets of varying number of markers and trios. We have found that the computational complexity of PHASE makes it prohibitive for routine use; on the other hand 2SNP, though the fastest method for small datasets, was significantly inaccurate. We have shown that our method outperforms BEAGLE in terms of speed and accuracy for small to intermediate dataset sizes in terms of number of trios for all marker sizes examined. Our method is implemented in the "Tree-Based Deterministic Sampling" (TDS package, available for download at http://www.ee.columbia.edu/~anastas/tds Conclusions Using a Tree-Based Deterministic sampling technique, we present an intuitive and conceptually simple phasing algorithm for trio data. The trade off between speed and accuracy achieved by our algorithm makes it a strong candidate for routine use on trio datasets.
An SQP Algorithm for Recourse-based Stochastic Nonlinear Programming
Xinshun Ma
2016-05-01
Full Text Available The stochastic nonlinear programming problem with completed recourse and nonlinear constraints is studied in this paper. We present a sequential quadratic programming method for solving the problem based on the certainty extended nonlinear model. This algorithm is obtained by combing the active set method and filter method. The convergence of the method is established under some standard assumptions. Moreover, a practical design is presented and numerical results are provided.
Adaptive algorithm for mobile user positioning based on environment estimation
Grujović Darko
2014-01-01
Full Text Available This paper analyzes the challenges to realize an infrastructure independent and a low-cost positioning method in cellular networks based on RSS (Received Signal Strength parameter, auxiliary timing parameter and environment estimation. The proposed algorithm has been evaluated using field measurements collected from GSM (Global System for Mobile Communications network, but it is technology independent and can be applied in UMTS (Universal Mobile Telecommunication Systems and LTE (Long-Term Evolution networks, also.
DELAUNAY-BASED SURFACE RECONSTRUCTION ALGORITHM IN REVERSE ENGINEERING
无
2002-01-01
Triangulation of scattered points is the first important section during reverse engineering. New concepts of dynamic circle and closed point are put forward based on current basic method. These new concepts can narrow the extent which triangulation process should seek through and optimize the triangles during producing them. Updating the searching edges dynamically controls progress of triangulation. Intersection judgment between new triangle and produced triangles is changed into intersection judgment between new triangle and searching edges. Examples illustrate superiorities of this new algorithm.
SENSITIVITY ANALYSIS BASED ON LANCZOS ALGORITHM IN STRUCTURAL DYNAMICS
李书; 王波; 胡继忠
2003-01-01
The sensitivity calculating formulas in structural dynamics was developed byutilizing the mathematical theorem and new definitions of sensitivities. So the singularityproblem of sensitivity with repeated eigenvalues is solved completely. To improve thecomputational efficiency, the reduction system is obtained based on Lanczos vectors. Afterincorporating the mathematical theory with the Lanczos algorithm, the approximatesensitivity solution can be obtained. A numerical example is presented to illustrate theperformance of the method.
Development of hybrid artificial intelligent based handover decision algorithm
A.M. Aibinu
2017-04-01
Full Text Available The possibility of seamless handover remains a mirage despite the plethora of existing handover algorithms. The underlying factor responsible for this has been traced to the Handover decision module in the Handover process. Hence, in this paper, the development of novel hybrid artificial intelligent handover decision algorithm has been developed. The developed model is made up of hybrid of Artificial Neural Network (ANN based prediction model and Fuzzy Logic. On accessing the network, the Received Signal Strength (RSS was acquired over a period of time to form a time series data. The data was then fed to the newly proposed k-step ahead ANN-based RSS prediction system for estimation of prediction model coefficients. The synaptic weights and adaptive coefficients of the trained ANN was then used to compute the k-step ahead ANN based RSS prediction model coefficients. The predicted RSS value was later codified as Fuzzy sets and in conjunction with other measured network parameters were fed into the Fuzzy logic controller in order to finalize handover decision process. The performance of the newly developed k-step ahead ANN based RSS prediction algorithm was evaluated using simulated and real data acquired from available mobile communication networks. Results obtained in both cases shows that the proposed algorithm is capable of predicting ahead the RSS value to about ±0.0002 dB. Also, the cascaded effect of the complete handover decision module was also evaluated. Results obtained show that the newly proposed hybrid approach was able to reduce ping-pong effect associated with other handover techniques.
A SAT-Based Algorithm for Context Matching
Bouquet, Paolo; Magnini, Bernardo; Serafini, Luciano; Zanobini, Stefano
2003-01-01
The development of more and more complex distributed applications over large networks of computers has raised the problem of semantic interoperability across applications based on local and autonomous semantic schemas (e.g., concept hierarchies, taxonomies, ontologies). In this paper we propose to view each semantic schema as a context (in the sense defined in \\cite{benerecetti9}), and propose an algorithm for automatically discovering relations across contexts. The main feature of the algori...
{WiFi GPS} based Combined positioning Algorithm
Zirari, Soumaya; Canalda, Philippe; Spies, François
2010-01-01
International audience; If nowadays, positioning becomes more and more accurate, and covers better and better a territory (indoor and outdoor), it remains territories where traditional (and basic) positioning system (GPS, gsm or WiFi) and hybrid ones (GPS-gsm, GPS-WiFi, GPS-WiFi-gsm,...) are insufficient and requires research investment treating combined positioning. In this paper we propose a GPS-WiFi combined positioning algorithm, based on trilateration technique. Real experiments and othe...
Physics-based signal processing algorithms for micromachined cantilever arrays
Candy, James V; Clague, David S; Lee, Christopher L; Rudd, Robert E; Burnham, Alan K; Tringe, Joseph W
2013-11-19
A method of using physics-based signal processing algorithms for micromachined cantilever arrays. The methods utilize deflection of a micromachined cantilever that represents the chemical, biological, or physical element being detected. One embodiment of the method comprises the steps of modeling the deflection of the micromachined cantilever producing a deflection model, sensing the deflection of the micromachined cantilever and producing a signal representing the deflection, and comparing the signal representing the deflection with the deflection model.
Support vector machines optimization based theory, algorithms, and extensions
Deng, Naiyang; Zhang, Chunhua
2013-01-01
Support Vector Machines: Optimization Based Theory, Algorithms, and Extensions presents an accessible treatment of the two main components of support vector machines (SVMs)-classification problems and regression problems. The book emphasizes the close connection between optimization theory and SVMs since optimization is one of the pillars on which SVMs are built.The authors share insight on many of their research achievements. They give a precise interpretation of statistical leaning theory for C-support vector classification. They also discuss regularized twi
Core Business Selection Based on Ant Colony Clustering Algorithm
Yu Lan; Yan Bo; Yao Baozhen
2014-01-01
Core business is the most important business to the enterprise in diversified business. In this paper, we first introduce the definition and characteristics of the core business and then descript the ant colony clustering algorithm. In order to test the effectiveness of the proposed method, Tianjin Port Logistics Development Co., Ltd. is selected as the research object. Based on the current situation of the development of the company, the core business of the company can be acquired by ant c...
NCUBE - A clustering algorithm based on a discretized data space
Eigen, D. J.; Northouse, R. A.
1974-01-01
Cluster analysis involves the unsupervised grouping of data. The process provides an automatic procedure for generating known training samples for pattern classification. NCUBE, the clustering algorithm presented, is based upon the concept of imposing a gridwork on the data space. The NCUBE computer implementation of this concept provides an easily derived form of piecewise linear discrimination. This piecewise linear discrimination permits the separation of some types of data groups that are not linearly separable.
An Index Based Skip Search Multiple Pattern Matching Algorithm
Raju Bhukya; Balram Parmer,; Anand Kulkarni
2011-01-01
DNA Pattern matching, the problem of finding sub sequences within a long DNA sequence has many applications in computational biology. As the sequences can be long, matching can be an expensive operation, especially as approximate matching is allowed. Searching DNA related data is a common activity for molecular biologists. In this paper we explore the applicability of a new pattern matching technique called Index based Skip Search Multiple Pattern matching algorithm (ISMPM), for DNA sequences...
DR-model-based estimation algorithm for NCS
HUANG Si-niu; CHEN Zong-ji; WEI Chen
2006-01-01
A novel estimation scheme based on dead reckoning (DR) model for networked control system (NCS)is proposed in this paper.Both the detailed DR estimation algorithm and the stability analysis of the system are given.By using DR estimation of the state,the effect of communication delays is overcome.This makes a controller designed without considering delays still applicable in NCS Moreover,the scheme can effectively solve the problem of data packet loss or timeout.
Model-based multiobjective evolutionary algorithm optimization for HCCI engines
Ma, He; Xu, Hongming; Wang, Jihong; Schnier, Thorsten; Neaves, Ben; Tan, Cheng; Wang, Zhi
2014-01-01
Modern engines feature a considerable number of adjustable control parameters. With this increasing number of Degrees of Freedom (DoF) for engines, and the consequent considerable calibration effort required to optimize engine performance, traditional manual engine calibration or optimization methods are reaching their limits. An automated engine optimization approach is desired. In this paper, a self-learning evolutionary algorithm based multi-objective globally optimization approach for a H...
Feature Selection for Image Retrieval based on Genetic Algorithm
Preeti Kushwaha
2016-12-01
Full Text Available This paper describes the development and implementation of feature selection for content based image retrieval. We are working on CBIR system with new efficient technique. In this system, we use multi feature extraction such as colour, texture and shape. The three techniques are used for feature extraction such as colour moment, gray level co- occurrence matrix and edge histogram descriptor. To reduce curse of dimensionality and find best optimal features from feature set using feature selection based on genetic algorithm. These features are divided into similar image classes using clustering for fast retrieval and improve the execution time. Clustering technique is done by k-means algorithm. The experimental result shows feature selection using GA reduces the time for retrieval and also increases the retrieval precision, thus it gives better and faster results as compared to normal image retrieval system. The result also shows precision and recall of proposed approach compared to previous approach for each image class. The CBIR system is more efficient and better performs using feature selection based on Genetic Algorithm.
Multifeature Fusion Vehicle Detection Algorithm Based on Choquet Integral
Wenhui Li
2014-01-01
Full Text Available Vision-based multivehicle detection plays an important role in Forward Collision Warning Systems (FCWS and Blind Spot Detection Systems (BSDS. The performance of these systems depends on the real-time capability, accuracy, and robustness of vehicle detection methods. To improve the accuracy of vehicle detection algorithm, we propose a multifeature fusion vehicle detection algorithm based on Choquet integral. This algorithm divides the vehicle detection problem into two phases: feature similarity measure and multifeature fusion. In the feature similarity measure phase, we first propose a taillight-based vehicle detection method, and then vehicle taillight feature similarity measure is defined. Second, combining with the definition of Choquet integral, the vehicle symmetry similarity measure and the HOG + AdaBoost feature similarity measure are defined. Finally, these three features are fused together by Choquet integral. Being evaluated on public test collections and our own test images, the experimental results show that our method has achieved effective and robust multivehicle detection in complicated environments. Our method can not only improve the detection rate but also reduce the false alarm rate, which meets the engineering requirements of Advanced Driving Assistance Systems (ADAS.
Detection of Human Head Direction Based on Facial Normal Algorithm
Lam Thanh Hien
2015-01-01
Full Text Available Many scholars worldwide have paid special efforts in searching for advance approaches to efficiently estimate human head direction which has been successfully applied in numerous applications such as human-computer interaction, teleconferencing, virtual reality, and 3D audio rendering. However, one of the existing shortcomings in the current literature is the violation of some ideal assumptions in practice. Hence, this paper aims at proposing a novel algorithm based on the normal of human face to recognize human head direction by optimizing a 3D face model combined with the facial normal model. In our experiments, a computational program was also developed based on the proposed algorithm and integrated with the surveillance system to alert the driver drowsiness. The program intakes data from either video or webcam, and then automatically identify the critical points of facial features based on the analysis of major components on the faces; and it keeps monitoring the slant angle of the head closely and makes alarming signal whenever the driver dozes off. From our empirical experiments, we found that our proposed algorithm effectively works in real-time basis and provides highly accurate results
Cryptographic protocol security analysis based on bounded constructing algorithm
无
2006-01-01
An efficient approach to analyzing cryptographic protocols is to develop automatic analysis tools based on formal methods. However, the approach has encountered the high computational complexity problem due to reasons that participants of protocols are arbitrary, their message structures are complex and their executions are concurrent. We propose an efficient automatic verifying algorithm for analyzing cryptographic protocols based on the Cryptographic Protocol Algebra (CPA) model proposed recently, in which algebraic techniques are used to simplify the description of cryptographic protocols and their executions. Redundant states generated in the analysis processes are much reduced by introducing a new algebraic technique called Universal Polynomial Equation and the algorithm can be used to verify the correctness of protocols in the infinite states space. We have implemented an efficient automatic analysis tool for cryptographic protocols, called ACT-SPA, based on this algorithm, and used the tool to check more than 20 cryptographic protocols. The analysis results show that this tool is more efficient, and an attack instance not offered previously is checked by using this tool.
ALGORITHMS FOR TENNIS RACKET ANALYSIS BASED ON MOTION DATA
Maria Skublewska-Paszkowska
2016-09-01
Full Text Available Modern technologies, such as motion capture systems (both optical and markerless, are more and more frequently used for athlete performance analysis due to their great precision. Optical systems based on the retro-reflective markers allow for tracking motion of multiple objects of various types. These systems compute human kinetic and kinematic parameters based on biomechanical models. Tracking additional objects like a tennis racket is also a very important aspect for analysing the player’s technique and precision. The motion data gathered by motion capture systems may be used for analysing various aspects that may not be recognised by the human eye or a video camera. This paper presents algorithms for analysis of a tennis racket motion during two of the most important tennis strokes: forehand and backhand. An optical Vicon system was used for obtaining the motion data which was the input for the algorithms. They indicate: the velocity of a tennis racket’s head and the racket’s handle based on the trajectories of attached markers as well as the racket’s orientation. The algorithms were implemented and tested on the data obtained from a professional trainer who participated in the research and performed a series of ten strikes, separately for: 1 forehand without a ball, 2 backhand without a ball, 3 forehand with a ball and 4 backhand with a ball. The computed parameters are gathered in tables and visualised in a graph.
Multiobjective immune algorithm with nondominated neighbor-based selection.
Gong, Maoguo; Jiao, Licheng; Du, Haifeng; Bo, Liefeng
2008-01-01
Abstract Nondominated Neighbor Immune Algorithm (NNIA) is proposed for multiobjective optimization by using a novel nondominated neighbor-based selection technique, an immune inspired operator, two heuristic search operators, and elitism. The unique selection technique of NNIA only selects minority isolated nondominated individuals in the population. The selected individuals are then cloned proportionally to their crowding-distance values before heuristic search. By using the nondominated neighbor-based selection and proportional cloning, NNIA pays more attention to the less-crowded regions of the current trade-off front. We compare NNIA with NSGA-II, SPEA2, PESA-II, and MISA in solving five DTLZ problems, five ZDT problems, and three low-dimensional problems. The statistical analysis based on three performance metrics including the coverage of two sets, the convergence metric, and the spacing, show that the unique selection method is effective, and NNIA is an effective algorithm for solving multiobjective optimization problems. The empirical study on NNIA's scalability with respect to the number of objectives shows that the new algorithm scales well along the number of objectives.
Genetic based optimization for multicast routing algorithm for MANET
C Rajan; N Shanthi
2015-12-01
Mobile Ad hoc Network (MANET) is established for a limited period, for special extemporaneous services related to mobile applications. This ad hoc network is set up for a limited period, in environments that change with the application. While in Internet the TCP/IP protocol suite supports a wide range of application, in MANETs protocols are tuned to specific customer/application. Multicasting is emerging as a popular communication format where the same packet is sent to multiple nodes in a network. Routing in multicasting involves maintaining routes and finding new node locations in a group and is NP-complete due to the dynamic nature of the network. In this paper, a Hybrid Genetic Based Optimization for Multicast Routing algorithm is proposed. The proposed algorithm uses the best features of Genetic Algorithm (GA) and particle swarm optimization (PSO) to improve the solution. Simulations were conducted by varying number of mobile nodes and results compared with Multicast AODV (MAODV) protocol, PSO based and GA based solution. The proposed optimization improves jitter, end to end delay and Packet Delivery Ratio (PDR) with faster convergence.
Wang, Youhua; Nakayama, Kenji
1995-01-01
In this letter, we introduce a predictor based least square (PLS) algorithm. By involving both order- and time-update recursions, the PLS algorithm is found to have a more stable performance compared with the stable version (Version II) of the RLS algorithm shown in Ref. [1]. Nevertheless, the computational requirement is about 50% of that of the RLS algorithm. As an application, the PLS algorithm can be applied to the fast newton transversal filters (FNTF) [2]. The FNTF algorithms suffer fro...
Genetic algorithm-based wide-band deterministic maximum likelihood direction finding algorithm
无
2005-01-01
The wide-band direction finding is one of hit and difficult task in array signal processing. This paper generalizes narrow-band deterministic maximum likelihood direction finding algorithm to the wideband case, and so constructions an object function, then utilizes genetic algorithm for nonlinear global optimization. Direction of arrival is estimated without preprocessing of array data and so the algorithm eliminates the effect of pre-estimate on the final estimation. The algorithm is applied on uniform linear array and extensive simulation results prove the efficacy of the algorithm. In the process of simulation, we obtain the relation between estimation error and parameters of genetic algorithm.
A Location-Based Business Information Recommendation Algorithm
Shudong Liu
2015-01-01
Full Text Available Recently, many researches on information (e.g., POI, ADs recommendation based on location have been done in both research and industry. In this paper, we firstly construct a region-based location graph (RLG, in which region node respectively connects with user node and business information node, and then we propose a location-based recommendation algorithm based on RLG, which can combine with user short-ranged mobility formed by daily activity and long-distance mobility formed by social network ties and sequentially can recommend local business information and long-distance business information to users. Moreover, it can combine user-based collaborative filtering with item-based collaborative filtering, and it can alleviate cold start problem which traditional recommender systems often suffer from. Empirical studies from large-scale real-world data from Yelp demonstrate that our method outperforms other methods on the aspect of recommendation accuracy.
Shao Wei; Qian Zuping; Yuan Feng
2007-01-01
A robust phase-only Direct Data Domain Least Squares (D3LS) algorithm based on generalized Rayleigh quotient optimization using hybrid Genetic Algorithm (GA) is presented in this letter. The optimization efficiency and computational speed are improved via the hybrid GA composed of standard GA and Nelder-Mead simplex algorithms. First, the objective function, with a form of generalized Rayleigh quotient, is derived via the standard D3LS algorithm. It is then taken as a fitness function and the unknown phases of all adaptive weights are taken as decision variables.Then, the nonlinear optimization is performed via the hybrid GA to obtain the optimized solution of phase-only adaptive weights. As a phase-only adaptive algorithm, the proposed algorithm is simpler than conventional algorithms when it comes to hardware implementation. Moreover, it processes only a single snapshot data as opposed to forming sample covariance matrix and operating matrix inversion. Simulation results show that the proposed algorithm has a good signal recovery and interferences nulling performance, which are superior to that of the phase-only D3LS algorithm based on standard GA.
Liu, Dong-sheng; Fan, Shu-jiang
2014-01-01
In order to offer mobile customers better service, we should classify the mobile user firstly. Aimed at the limitations of previous classification methods, this paper puts forward a modified decision tree algorithm for mobile user classification, which introduced genetic algorithm to optimize the results of the decision tree algorithm. We also take the context information as a classification attributes for the mobile user and we classify the context into public context and private context classes. Then we analyze the processes and operators of the algorithm. At last, we make an experiment on the mobile user with the algorithm, we can classify the mobile user into Basic service user, E-service user, Plus service user, and Total service user classes and we can also get some rules about the mobile user. Compared to C4.5 decision tree algorithm and SVM algorithm, the algorithm we proposed in this paper has higher accuracy and more simplicity.
An Enhanced Differential Evolution Algorithm Based on Multiple Mutation Strategies
Wan-li Xiang
2015-01-01
Full Text Available Differential evolution algorithm is a simple yet efficient metaheuristic for global optimization over continuous spaces. However, there is a shortcoming of premature convergence in standard DE, especially in DE/best/1/bin. In order to take advantage of direction guidance information of the best individual of DE/best/1/bin and avoid getting into local trap, based on multiple mutation strategies, an enhanced differential evolution algorithm, named EDE, is proposed in this paper. In the EDE algorithm, an initialization technique, opposition-based learning initialization for improving the initial solution quality, and a new combined mutation strategy composed of DE/current/1/bin together with DE/pbest/bin/1 for the sake of accelerating standard DE and preventing DE from clustering around the global best individual, as well as a perturbation scheme for further avoiding premature convergence, are integrated. In addition, we also introduce two linear time-varying functions, which are used to decide which solution search equation is chosen at the phases of mutation and perturbation, respectively. Experimental results tested on twenty-five benchmark functions show that EDE is far better than the standard DE. In further comparisons, EDE is compared with other five state-of-the-art approaches and related results show that EDE is still superior to or at least equal to these methods on most of benchmark functions.
A meta-learning system based on genetic algorithms
Pellerin, Eric; Pigeon, Luc; Delisle, Sylvain
2004-04-01
The design of an efficient machine learning process through self-adaptation is a great challenge. The goal of meta-learning is to build a self-adaptive learning system that is constantly adapting to its specific (and dynamic) environment. To that end, the meta-learning mechanism must improve its bias dynamically by updating the current learning strategy in accordance with its available experiences or meta-knowledge. We suggest using genetic algorithms as the basis of an adaptive system. In this work, we propose a meta-learning system based on a combination of the a priori and a posteriori concepts. A priori refers to input information and knowledge available at the beginning in order to built and evolve one or more sets of parameters by exploiting the context of the system"s information. The self-learning component is based on genetic algorithms and neural Darwinism. A posteriori refers to the implicit knowledge discovered by estimation of the future states of parameters and is also applied to the finding of optimal parameters values. The in-progress research presented here suggests a framework for the discovery of knowledge that can support human experts in their intelligence information assessment tasks. The conclusion presents avenues for further research in genetic algorithms and their capability to learn to learn.
Formal Photograph Compression Algorithm Based on Object Segmentation
Li Zhu; Guo-You Wang; Chen Wang
2008-01-01
Small storage space for photographs in formal documents is increasingly necessary in today's needs for huge amounts of data communication and storage. Traditional compression algorithms do not sufficiently utilize the distinctness of formal photographs. That is, the object is an image of the human head, and the background is in unicolor. Therefore, the compression is of low efficiency and the image after compression is still space-consuming. This paper presents an image compression algorithm based on object segmentation for practical high-efficiency applications. To achieve high coding efficiency, shape-adaptive discrete wavelet transforms are used to transformation arbitrarily shaped objects. The areas of the human head and its background are compressed separately to reduce the coding redundancy of the background. Two methods, lossless image contour coding based on differential chain, and modified set partitioning in hierarchical trees (SPIHT) algorithm of arbitrary shape, are discussed in detail. The results of experiments show that when bit per pixel (bpp)is equal to 0.078, peak signal-to-noise ratio (PSNR) of reconstructed photograph will exceed the standard of SPIHT by nearly 4dB.
A virtual network mapping algorithm based on integer programming
Bo LU; Jian-ya CHEN; Hong-yan CUI; Tao HUANG; Yun-jie LIU
2013-01-01
The virtual network (VN) embedding/mapping problem is recognized as an essential question of network virtualiza-tion. The VN embedding problem is a major challenge in this field. Its target is to efficiently map the virtual nodes and virtual links onto the substrate network resources. Previous research focused on designing heuristic-based algorithms or attempting two-stage solutions by solving node mapping in the first stage and link mapping in the second stage. In this study, we propose a new VN embedding algorithm based on integer programming. We build a model of an augmented substrate graph, and formulate the VN embedding problem as an integer program with an objective function and some constraints. A factor of topology-awareness is added to the objective function. The VN embedding problem is solved in one stage. Simulation results show that our algorithm greatly enhances the acceptance ratio, and increases the revenue/cost (R/C) ratio and the revenue while decreasing the cost of the VN embedding problem.
A Radio-Map Automatic Construction Algorithm Based on Crowdsourcing
Yu, Ning; Xiao, Chenxian; Wu, Yinfeng; Feng, Renjian
2016-01-01
Traditional radio-map-based localization methods need to sample a large number of location fingerprints offline, which requires huge amount of human and material resources. To solve the high sampling cost problem, an automatic radio-map construction algorithm based on crowdsourcing is proposed. The algorithm employs the crowd-sourced information provided by a large number of users when they are walking in the buildings as the source of location fingerprint data. Through the variation characteristics of users’ smartphone sensors, the indoor anchors (doors) are identified and their locations are regarded as reference positions of the whole radio-map. The AP-Cluster method is used to cluster the crowdsourced fingerprints to acquire the representative fingerprints. According to the reference positions and the similarity between fingerprints, the representative fingerprints are linked to their corresponding physical locations and the radio-map is generated. Experimental results demonstrate that the proposed algorithm reduces the cost of fingerprint sampling and radio-map construction and guarantees the localization accuracy. The proposed method does not require users’ explicit participation, which effectively solves the resource-consumption problem when a location fingerprint database is established. PMID:27070623
A Radio-Map Automatic Construction Algorithm Based on Crowdsourcing
Ning Yu
2016-04-01
Full Text Available Traditional radio-map-based localization methods need to sample a large number of location fingerprints offline, which requires huge amount of human and material resources. To solve the high sampling cost problem, an automatic radio-map construction algorithm based on crowdsourcing is proposed. The algorithm employs the crowd-sourced information provided by a large number of users when they are walking in the buildings as the source of location fingerprint data. Through the variation characteristics of users’ smartphone sensors, the indoor anchors (doors are identified and their locations are regarded as reference positions of the whole radio-map. The AP-Cluster method is used to cluster the crowdsourced fingerprints to acquire the representative fingerprints. According to the reference positions and the similarity between fingerprints, the representative fingerprints are linked to their corresponding physical locations and the radio-map is generated. Experimental results demonstrate that the proposed algorithm reduces the cost of fingerprint sampling and radio-map construction and guarantees the localization accuracy. The proposed method does not require users’ explicit participation, which effectively solves the resource-consumption problem when a location fingerprint database is established.
Matched field localization based on CS-MUSIC algorithm
Guo, Shuangle; Tang, Ruichun; Peng, Linhui; Ji, Xiaopeng
2016-04-01
The problem caused by shortness or excessiveness of snapshots and by coherent sources in underwater acoustic positioning is considered. A matched field localization algorithm based on CS-MUSIC (Compressive Sensing Multiple Signal Classification) is proposed based on the sparse mathematical model of the underwater positioning. The signal matrix is calculated through the SVD (Singular Value Decomposition) of the observation matrix. The observation matrix in the sparse mathematical model is replaced by the signal matrix, and a new concise sparse mathematical model is obtained, which means not only the scale of the localization problem but also the noise level is reduced; then the new sparse mathematical model is solved by the CS-MUSIC algorithm which is a combination of CS (Compressive Sensing) method and MUSIC (Multiple Signal Classification) method. The algorithm proposed in this paper can overcome effectively the difficulties caused by correlated sources and shortness of snapshots, and it can also reduce the time complexity and noise level of the localization problem by using the SVD of the observation matrix when the number of snapshots is large, which will be proved in this paper.
Quadratic Error Metric Mesh Simplification Algorithm Based on Discrete Curvature
Li Yao
2015-01-01
Full Text Available Complex and highly detailed polygon meshes have been adopted for model representation in many areas of computer graphics. Existing works mainly focused on the quadric error metric based complex models approximation, which has not taken the retention of important model details into account. This may lead to visual degeneration. In this paper, we improve Garland and Heckberts’ quadric error metric based algorithm by using the discrete curvature to reserve more features for mesh simplification. Our experiments on various models show that the geometry and topology structure as well as the features of the original models are precisely retained by employing discrete curvature.
Research of Video Steganalysis Algorithm Based on H265 Protocol
Wu Kaicheng
2015-01-01
This paper researches LSB matching VSA based on H265 protocol with the research background of 26 original Video sequences, it firstly extracts classification features out from training samples as input of SVM, and trains in SVM to obtain high-quality category classification model, and then tests whether there is suspicious information in the video sample. The experimental results show that VSA algorithm based on LSB matching can be more practical to obtain all frame embedded secret information and carrier and video of local frame embedded. In addition, VSA adopts the method of frame by frame with a strong robustness in resisting attack in the corresponding time domain.
Genetic Algorithm based Decentralized PI Type Controller: Load Frequency Control
Dwivedi, Atul; Ray, Goshaidas; Sharma, Arun Kumar
2016-12-01
This work presents a design of decentralized PI type Linear Quadratic (LQ) controller based on genetic algorithm (GA). The proposed design technique allows considerable flexibility in defining the control objectives and it does not consider any knowledge of the system matrices and moreover it avoids the solution of algebraic Riccati equation. To illustrate the results of this work, a load-frequency control problem is considered. Simulation results reveal that the proposed scheme based on GA is an alternative and attractive approach to solve load-frequency control problem from both performance and design point of views.
3D face recognition algorithm based on detecting reliable components
Huang Wenjun; Zhou Xuebing; Niu Xiamu
2007-01-01
Fisherfaces algorithm is a popular method for face recognition. However, there exist some unstable components that degrade recognition performance. In this paper, we propose a method based on detecting reliable components to overcome the problem and introduce it to 3D face recognition. The reliable components are detected within the binary feature vector, which is generated from the Fisherfaces feature vector based on statistical properties, and is used for 3D face recognition as the final feature vector. Experimental results show that the reliable components feature vector is much more effective than the Fisherfaces feature vector for face recognition.
The algorithm of malicious code detection based on data mining
Yang, Yubo; Zhao, Yang; Liu, Xiabi
2017-08-01
Traditional technology of malicious code detection has low accuracy and it has insufficient detection capability for new variants. In terms of malicious code detection technology which is based on the data mining, its indicators are not accurate enough, and its classification detection efficiency is relatively low. This paper proposed the information gain ratio indicator based on the N-gram to choose signature, this indicator can accurately reflect the detection weight of the signature, and helped by C4.5 decision tree to elevate the algorithm of classification detection.
Segment-based traffic smoothing algorithm for VBR video stream
无
2006-01-01
Transmission of variable bit rate (VBR) video, because of the burstiness of VBR video traffic, has high fluctuation in bandwidth requirement. Traffic smoothing algorithm is very efficient in reducing burstiness of the VBR video stream by transmitting data in a series of fixed rates. We propose in this paper a novel segment-based bandwidth allocation algorithm which dynamically adjusts the segmentation boundary and changes the transmission rate at the latest possible point so that the video segment will be extended as long as possible and the number of rate changes can be as small as possible while keeping the peak rate low. Simulation results showed that our approach has small bandwidth requirement, high bandwidth utilization and low computation cost.
An Improved Piecewise Linear Chaotic Map Based Image Encryption Algorithm
Yuping Hu
2014-01-01
Full Text Available An image encryption algorithm based on improved piecewise linear chaotic map (MPWLCM model was proposed. The algorithm uses the MPWLCM to permute and diffuse plain image simultaneously. Due to the sensitivity to initial key values, system parameters, and ergodicity in chaotic system, two pseudorandom sequences are designed and used in the processes of permutation and diffusion. The order of processing pixels is not in accordance with the index of pixels, but it is from beginning or end alternately. The cipher feedback was introduced in diffusion process. Test results and security analysis show that not only the scheme can achieve good encryption results but also its key space is large enough to resist against brute attack.
Heuristic Based Adaptive Step Size CLMS Algorithms for Smart Antennas
Y Rama Krishna
2013-05-01
Full Text Available A smart antenna system combines multiple antenna elements with a signal processing capability to optimize its radiation and/or reception pattern automatically in response to the signal environment through complex weight selection. The weight selection process to get suitable Array factor with low Half Power Beam Width (HPBW and Side Lobe Level (SLL is a complex method. The aim of this task is to design a new approach for smart antennas to minimize the noise and interference effects from external sources with least number of iterations. This paper presents Heuristics based adaptive step size Complex Least Mean Square (CLMS model for Smart Antennas to speedup convergence. In this process Benveniste and Mathews algorithms are used as heuristics with CLMS and the improvement of performance of Smart Antenna System in terms of convergence rate and array factor are discussed and compared with the performance of CLMS and Augmented CLMS (ACLMS algorithms.
A geometric reasoning based algorithm for point pattern matching
徐文立; 张立华
2001-01-01
Point pattern matching (PPM) is an important topic in computer vision and pattern recognition. It can be widely used in many areas such as image registration, object recognition, motion detection, target tracking, autonomous navigation, and pose estimation. This paper discusses the incomplete matching problem of two point sets under Euclidean transformation. According to geometric reasoning, some definitions for matching clique, support point pair, support index set, and support index matrix, etc. are given. Based on the properties and theorems of them, a novel reasoning algorithm is presented, which searches for the optimal sOlLtion from top to bottom and could find out as many consistent corresponding point pairs as possible. Theoretical analysis and experimental results show that the new algorithm is very effective, and could be, under some conditions, applied to the PPM problem under other kind of transformations.
ROBIL: Robot Path Planning Based on PBIL Algorithm
Bo-Yeong Kang
2014-09-01
Full Text Available Genetic algorithm (GAs have attracted considerable interest for their usefulness in solving complex robot path planning problems. Specifically, researchers have combined conventional GAs with problem-specific operators and initialization techniques to find the shortest paths in a variety of robotic environments. Unfortunately, these approaches have exhibited inherently unstable performance, and they have tended to make other aspects of the problem-solving process (e.g., adjusting parameter sensitivities and creating high-quality initial populations unmanageable. As an alternative to conventional GAs, we propose a new population-based incremental learning (PBIL algorithm for robot path planning, a probabilistic model of nodes, and an edge bank for generating promising paths. Experimental results demonstrate the computational superiority of the proposed method over conventional GA approaches.
Indexing Algorithm Based on Improved Sparse Local Sensitive Hashing
Yiwei Zhu
2014-01-01
Full Text Available In this article, we propose a new semantic hashing algorithm to address the new-merging problems such as the difficulty in similarity measurement brought by high-dimensional data. Based on local sensitive hashing and spectral hashing, we introduce sparse principal component analysis (SPCA to reduce the dimension of the data set which exclude the redundancy in the parameter list, and thus make high dimensional indexing and retrieval faster and more efficient. In the meanwhile, we employ Boosting algorithm in machine learning to determine the threshold of hashing, so as to improve its adaptive ability to real data and extend its range of application. According to experiments, this method not only has satisfying performance on multimedia data sets such as images and texts, but also performs better than the common indexing methods.
Printing Detecting Algorithm Basing on Maximum Degree of Recognition
Hu Zhang
2013-04-01
Full Text Available In modern packaging, printing industry, due to effects of the properties of the strip itself and the ambient light, strip background color and the color of the printing line, the low contrast boundaries of the strip on both sides and so on, the traditional digital qualitative detection and control to the correction system does not meet the comprehensive requirements. This paper aims to study the detection of a continuous line, discontinuous line and color dividing line on the strip, and because of low contrast between background color and dividing line, we proposed an innovative solution and implementation. This article discusses a new algorithm basing on maximum degree of recognition and optimal light source search algorithm, and we simulated this in MATLAB, finally, we completed the physical testing of the overall system.
Remote Sensing Image Resolution Enlargement Algorithm Based on Wavelet Transformation
Samiul Azam
2014-05-01
Full Text Available In this paper, we present a new image resolution enhancement algorithm based on cycle spinning and stationary wavelet subband padding. The proposed technique or algorithm uses stationary wavelet transformation (SWT to decompose the low resolution (LR image into frequency subbands. All these frequency subbands are interpolated using either bicubic or lanczos interpolation, and these interpolated subbands are put into inverse SWT process for generating intermediate high resolution (HR image. Finally, cycle spinning (CS is applied on this intermediate high resolution image for reducing blocking artifacts, followed by, traditional Laplacian sharpening filter is used to make the generated high resolution image sharper. This new technique has been tested on several satellite images. Experimental result shows that the proposed technique outperforms the conventional and the state-of-the-art techniques in terms of peak signal to noise ratio, root mean square error, entropy, as well as, visual perspective.
Machine learning based global particle indentification algorithms at LHCb experiment
Derkach, Denis; Likhomanenko, Tatiana; Rogozhnikov, Aleksei; Ratnikov, Fedor
2017-01-01
One of the most important aspects of data processing at LHC experiments is the particle identification (PID) algorithm. In LHCb, several different sub-detector systems provide PID information: the Ring Imaging CHerenkov (RICH) detector, the hadronic and electromagnetic calorimeters, and the muon chambers. To improve charged particle identification, several neural networks including a deep architecture and gradient boosting have been applied to data. These new approaches provide higher identification efficiencies than existing implementations for all charged particle types. It is also necessary to achieve a flat dependency between efficiencies and spectator variables such as particle momentum, in order to reduce systematic uncertainties during later stages of data analysis. For this purpose, "flat” algorithms that guarantee the flatness property for efficiencies have also been developed. This talk presents this new approach based on machine learning and its performance.
Adaptive inpainting algorithm based on DCT induced wavelet regularization.
Li, Yan-Ran; Shen, Lixin; Suter, Bruce W
2013-02-01
In this paper, we propose an image inpainting optimization model whose objective function is a smoothed l(1) norm of the weighted nondecimated discrete cosine transform (DCT) coefficients of the underlying image. By identifying the objective function of the proposed model as a sum of a differentiable term and a nondifferentiable term, we present a basic algorithm inspired by Beck and Teboulle's recent work on the model. Based on this basic algorithm, we propose an automatic way to determine the weights involved in the model and update them in each iteration. The DCT as an orthogonal transform is used in various applications. We view the rows of a DCT matrix as the filters associated with a multiresolution analysis. Nondecimated wavelet transforms with these filters are explored in order to analyze the images to be inpainted. Our numerical experiments verify that under the proposed framework, the filters from a DCT matrix demonstrate promise for the task of image inpainting.
Multirobot FastSLAM Algorithm Based on Landmark Consistency Correction
Shi-Ming Chen
2014-01-01
Full Text Available Considering the influence of uncertain map information on multirobot SLAM problem, a multirobot FastSLAM algorithm based on landmark consistency correction is proposed. Firstly, electromagnetism-like mechanism is introduced to the resampling procedure in single-robot FastSLAM, where we assume that each sampling particle is looked at as a charged electron and attraction-repulsion mechanism in electromagnetism field is used to simulate interactive force between the particles to improve the distribution of particles. Secondly, when multiple robots observe the same landmarks, every robot is regarded as one node and Kalman-Consensus Filter is proposed to update landmark information, which further improves the accuracy of localization and mapping. Finally, the simulation results show that the algorithm is suitable and effective.
Research on Wavelet-Based Algorithm for Image Contrast Enhancement
Wu Ying-qian; Du Pei-jun; Shi Peng-fei
2004-01-01
A novel wavelet-based algorithm for image enhancement is proposed in the paper. On the basis of multiscale analysis, the proposed algorithm solves efficiently the problem of noise over-enhancement, which commonly occurs in the traditional methods for contrast enhancement. The decomposed coefficients at same scales are processed by a nonlinear method, and the coefficients at different scales are enhanced in different degree. During the procedure, the method takes full advantage of the properties of Human visual system so as to achieve better performance. The simulations demonstrate that these characters of the proposed approach enable it to fully enhance the content in images, to efficiently alleviate the enhancement of noise and to achieve much better enhancement effect than the traditional approaches.
Routing Optimization Based on Taboo Search Algorithm for Logistic Distribution
Hongxue Yang
2014-04-01
Full Text Available Along with the widespread application of the electronic commerce in the modern business, the logistic distribution has become increasingly important. More and more enterprises recognize that the logistic distribution plays an important role in the process of production and sales. A good routing for logistic distribution can cut down transport cost and improve efficiency. In order to cut down transport cost and improve efficiency, a routing optimization based on taboo search for logistic distribution is proposed in this paper. Taboo search is a metaheuristic search method to perform local search used for logistic optimization. The taboo search is employed to accelerate convergence and the aspiration criterion is combined with the heuristics algorithm to solve routing optimization. Simulation experimental results demonstrate that the optimal routing in the logistic distribution can be quickly obtained by the taboo search algorithm
Cluster-Based Distributed Algorithms for Very Large Linear Equations
无
2006-01-01
In many applications such as computational fluid dynamics and weather prediction, as well as image processing and state of Markov chain etc., the grade of matrix n is often very large, and any serial algorithm cannot solve the problems. A distributed cluster-based solution for very large linear equations is discussed, it includes the definitions of notations, partition of matrix, communication mechanism, and a master-slaver algorithm etc., the computing cost is O(n3/N), the memory cost is O(n2/N), the I/O cost is O(n2/N), and the communication cost is O(Nn), here, N is the number of computing nodes or processes. Some tests show that the solution could solve the double type of matrix under 106×106 effectively.
A NEW RSA CRYPTOSYSTEM HARDWARE IMPLEMENTATION BASED ON MONTGOMERY'S ALGORITHM
卢君明; 林争辉
2002-01-01
RSA public key crypto-system is a relatively safe technology, which is widely used in today's secure electronic communication. In this paper, a new implementation method to optimize a 1 024 bit RSA processor was presented. Basically, a fast modular multiplication architecture based on Montgomery's algorithm was proposed. Modular exponentiation algorithm scans encryption from right to left, so two modular multiplications can be processed parallel. The new architecture is also fit for an effective I/O interface. The time to calculate a modular exponentiation is about n2 clock cycles. The proposed architecture has a data rate of 93.7 kb/s for 1 024 bit work with a 100 MHz clock.
An Improved Piecewise Linear Chaotic Map Based Image Encryption Algorithm
Hu, Yuping; Wang, Zhijian
2014-01-01
An image encryption algorithm based on improved piecewise linear chaotic map (MPWLCM) model was proposed. The algorithm uses the MPWLCM to permute and diffuse plain image simultaneously. Due to the sensitivity to initial key values, system parameters, and ergodicity in chaotic system, two pseudorandom sequences are designed and used in the processes of permutation and diffusion. The order of processing pixels is not in accordance with the index of pixels, but it is from beginning or end alternately. The cipher feedback was introduced in diffusion process. Test results and security analysis show that not only the scheme can achieve good encryption results but also its key space is large enough to resist against brute attack. PMID:24592159
A Matrix Pencil Algorithm Based Multiband Iterative Fusion Imaging Method
Zou, Yong Qiang; Gao, Xun Zhang; Li, Xiang; Liu, Yong Xiang
2016-01-01
Multiband signal fusion technique is a practicable and efficient way to improve the range resolution of ISAR image. The classical fusion method estimates the poles of each subband signal by the root-MUSIC method, and some good results were get in several experiments. However, this method is fragile in noise for the proper poles could not easy to get in low signal to noise ratio (SNR). In order to eliminate the influence of noise, this paper propose a matrix pencil algorithm based method to estimate the multiband signal poles. And to deal with mutual incoherent between subband signals, the incoherent parameters (ICP) are predicted through the relation of corresponding poles of each subband. Then, an iterative algorithm which aimed to minimize the 2-norm of signal difference is introduced to reduce signal fusion error. Applications to simulate dada verify that the proposed method get better fusion results at low SNR.
Autonomous Navigation of Mobile Robot Based on Flood Fill Algorithm
Ayad Mohammed Jabbar
2016-06-01
Full Text Available The autonomous navigation of robots is an important area of research. It can intelligently navigate itself from source to target within an environment without human interaction. Recently, algorithms and techniques have been made and developed to improve the performance of robots. It’s more effective and has high precision tasks than before. This work proposed to solve a maze using a Flood fill algorithm based on real time camera monitoring the movement on its environment. Live video streaming sends an obtained data to be processed by the server. The server sends back the information to the robot via wireless radio. The robot works as a client device moves from point to point depends on server information. Using camera in this work allows voiding great time that needs it to indicate the route by the robot.
Healing Temperature of Hybrid Structures Based on Genetic Algorithm
赵中伟; 陈志华; 刘红波
2016-01-01
The healing temperature of suspen-dome with stacked arches(SDSA)and arch-supported single-layer lattice shell structures was investigated based on the genetic algorithm. The temperature field of arch under solar radiation was derived by FLUENT to investigate the influence of solar radiation on the determination of the healing temperature. Moreover, a multi-scale model was established to apply the complex temperature field under solar radiation. The change in the mechanical response of these two kinds of structures with the healing temperature was discussed. It can be concluded that solar radiation has great influence on the healing temperature, and the genetic algorithm can be effectively used in the optimization of the healing temperature for hybrid structures.
Complete Boolean Satisfiability Solving Algorithms Based on Local Search
Wen-Sheng Guo; Guo-Wu Yang; William N.N.Hung; Xiaoyu Song
2013-01-01
Boolean satisfiability (SAT) is a well-known problem in computer science,artificial intelligence,and operations research.This paper focuses on the satisfiability problem of Model RB structure that is similar to graph coloring problems and others.We propose a translation method and three effective complete SAT solving algorithms based on the characterization of Model RB structure.We translate clauses into a graph with exclusive sets and relative sets.In order to reduce search depth,we determine search order using vertex weights and clique in the graph.The results show that our algorithms are much more effective than the best SAT solvers in numerous Model RB benchmarks,especially in those large benchmark instances.
FPGA-Based Implementation of Lithuanian Isolated Word Recognition Algorithm
Tomyslav Sledevič
2013-05-01
Full Text Available The paper describes the FPGA-based implementation of Lithuanian isolated word recognition algorithm. FPGA is selected for parallel process implementation using VHDL to ensure fast signal processing at low rate clock signal. Cepstrum analysis was applied to features extraction in voice. The dynamic time warping algorithm was used to compare the vectors of cepstrum coefficients. A library of 100 words features was created and stored in the internal FPGA BRAM memory. Experimental testing with speaker dependent records demonstrated the recognition rate of 94%. The recognition rate of 58% was achieved for speaker-independent records. Calculation of cepstrum coefficients lasted for 8.52 ms at 50 MHz clock, while 100 DTWs took 66.56 ms at 25 MHz clock.Article in Lithuanian
An OFDMA resource allocation algorithm based on coalitional games
Bacci Giacomo
2011-01-01
Full Text Available Abstract This work investigates a fair adaptive resource management criterion (in terms of transmit powers and subchannel assignment for the uplink of an orthogonal frequency-division multiple access network, populated by mobile users with constraints in terms of target data rates. The inherent optimization problem is tackled with the analytical tools of coalitional game theory, and a practical algorithm based on Markov modeling is introduced. The proposed scheme allows the mobile devices to fulfill their rate demands exactly with a minimum utilization of network resources. Simulation results show that the average number of operations of the proposed iterative algorithm are much lower than K · N, where N and K are the number of allocated subcarriers and of mobile terminals.
Zhao Ruizhen; Ren Xiaoxin; Han Xuelian; Hu Shaohai
2012-01-01
Sparsity Adaptive Matching Pursuit (SAMP) algorithm is a widely used reconstruction algorithm for compressive sensing in the case that the sparsity is unknown.In order to match the sparsity more accurately,we presented an improved SAMP algorithm based on Regularized Backtracking (SAMP-RB).By adapting a regularized backtracking step to SAMP algorithm in each iteration stage,the proposed algorithm can flexibly remove the inappropriate atoms.The experimental results show that SAMP-RB reconstruction algorithm greatly improves SAMP algorithm both in reconstruction quality and computational time.It has better reconstruction efficiency than most of the available matching pursuit algorithms.
A Block-Based Multi-Scale Background Extraction Algorithm
Seyed H. Davarpanah
2010-01-01
Full Text Available Problem statement: To extract the moving objects, vision-based surveillance systems subtract the current image from a predefined background image. The efficiency of these systems mainly depends on accuracy of the extracted background image. It should be able to adapt to the changes continuously. In addition, especially in real-time applications the time complexity of this adaptation is a critical matter. Approach: In this study, to extract an adaptive background, a combination of blocking and multi-scale methods is presented. Because of being less sensitive to local movements, block-based techniques are proper to control the non-stationary objects movements, especially in outdoor applications. They can be useful to reduce the effect of these objects on the extracted background. We also used the blocking method to intelligently select the regions which the temporal filtering has to be applied on. In addition, an amended multi-scale algorithm is introduced. This algorithm is a hybrid algorithm, a combination of some nonparametric and parametric filters. It uses a nonparametric filter in the spatial domain to initiate two primary backgrounds. In continue two adapted two-dimensional filters will be used to extract the final background. Results: The qualitative and quantitative results of our experiments certify not only the quality of the final extracted background is acceptable, but also its time consumption is approximately half in compare to the similar methods. Conclusion: Using Multi scaling filtering and applying the filters just to some selected nonoverlapped blocks reduce the time consumption of the extracting background algorithm.
Algorithmic support for commodity-based parallel computing systems.
Leung, Vitus Joseph; Bender, Michael A. (State University of New York, Stony Brook, NY); Bunde, David P. (University of Illinois, Urbna, IL); Phillips, Cynthia Ann
2003-10-01
The Computational Plant or Cplant is a commodity-based distributed-memory supercomputer under development at Sandia National Laboratories. Distributed-memory supercomputers run many parallel programs simultaneously. Users submit their programs to a job queue. When a job is scheduled to run, it is assigned to a set of available processors. Job runtime depends not only on the number of processors but also on the particular set of processors assigned to it. Jobs should be allocated to localized clusters of processors to minimize communication costs and to avoid bandwidth contention caused by overlapping jobs. This report introduces new allocation strategies and performance metrics based on space-filling curves and one dimensional allocation strategies. These algorithms are general and simple. Preliminary simulations and Cplant experiments indicate that both space-filling curves and one-dimensional packing improve processor locality compared to the sorted free list strategy previously used on Cplant. These new allocation strategies are implemented in Release 2.0 of the Cplant System Software that was phased into the Cplant systems at Sandia by May 2002. Experimental results then demonstrated that the average number of communication hops between the processors allocated to a job strongly correlates with the job's completion time. This report also gives processor-allocation algorithms for minimizing the average number of communication hops between the assigned processors for grid architectures. The associated clustering problem is as follows: Given n points in {Re}d, find k points that minimize their average pairwise L{sub 1} distance. Exact and approximate algorithms are given for these optimization problems. One of these algorithms has been implemented on Cplant and will be included in Cplant System Software, Version 2.1, to be released. In more preliminary work, we suggest improvements to the scheduler separate from the allocator.
Fan, Desheng; Meng, Xiangfeng; Wang, Yurong; Yang, Xiulun; Peng, Xiang; He, Wenqi; Dong, Guoyan; Chen, Hongyi
2013-08-10
An optical identity authentication scheme based on the elliptic curve digital signature algorithm (ECDSA) and phase retrieval algorithm (PRA) is proposed. In this scheme, a user's certification image and the quick response code of the user identity's keyed-hash message authentication code (HMAC) with added noise, serving as the amplitude and phase restriction, respectively, are digitally encoded into two phase keys using a PRA in the Fresnel domain. During the authentication process, when the two phase keys are presented to the system and illuminated by a plane wave of correct wavelength, an output image is generated in the output plane. By identifying whether there is a match between the amplitude of the output image and all the certification images pre-stored in the database, the system can thus accomplish a first-level verification. After the confirmation of first-level verification, the ECDSA signature is decoded from the phase part of the output image and verified to allege whether the user's identity is legal or not. Moreover, the introduction of HMAC makes it almost impossible to forge the signature and hence the phase keys thanks to the HMAC's irreversible property. Theoretical analysis and numerical simulations both validate the feasibility of our proposed scheme.
Vision Based Autonomous Robot Navigation Algorithms and Implementations
Chatterjee, Amitava; Nirmal Singh, N
2013-01-01
This book is devoted to the theory and development of autonomous navigation of mobile robots using computer vision based sensing mechanism. The conventional robot navigation systems, utilizing traditional sensors like ultrasonic, IR, GPS, laser sensors etc., suffer several drawbacks related to either the physical limitations of the sensor or incur high cost. Vision sensing has emerged as a popular alternative where cameras can be used to reduce the overall cost, maintaining high degree of intelligence, flexibility and robustness. This book includes a detailed description of several new approaches for real life vision based autonomous navigation algorithms and SLAM. It presents the concept of how subgoal based goal-driven navigation can be carried out using vision sensing. The development concept of vision based robots for path/line tracking using fuzzy logic is presented, as well as how a low-cost robot can be indigenously developed in the laboratory with microcontroller based sensor systems. The book descri...
Electromagnetic Model and Image Reconstruction Algorithms Based on EIT System
CAO Zhang; WANG Huaxiang
2006-01-01
An intuitive 2 D model of circular electrical impedance tomography ( EIT) sensor with small size electrodes is established based on the theory of analytic functions.The validation of the model is proved using the result from the solution of Laplace equation.Suggestions on to electrode optimization and explanation to the ill-condition property of the sensitivity matrix are provided based on the model,which takes electrode distance into account and can be generalized to the sensor with any simple connected region through a conformal transformation.Image reconstruction algorithms based on the model are implemented to show feasibility of the model using experimental data collected from the EIT system developed in Tianjin University.In the simulation with a human chestlike configuration,electrical conductivity distributions are reconstructed using equi-potential backprojection (EBP) and Tikhonov regularization (TR) based on a conformal transformation of the model.The algorithms based on the model are suitable for online image reconstruction and the reconstructed results are good both in size and position.
Tugrul Talaslioglu
2009-01-01
Full Text Available A new genetic algorithm (GA methodology, Bipopulation-Based Genetic Algorithm with Enhanced Interval Search (BGAwEIS, is introduced and used to optimize the design of truss structures with various complexities. The results of BGAwEIS are compared with those obtained by the sequential genetic algorithm (SGA utilizing a single population, a multipopulation-based genetic algorithm (MPGA proposed for this study and other existing approaches presented in literature. This study has two goals: outlining BGAwEIS's fundamentals and evaluating the performances of BGAwEIS and MPGA. Consequently, it is demonstrated that MPGA shows a better performance than SGA taking advantage of multiple populations, but BGAwEIS explores promising solution regions more efficiently than MPGA by exploiting the feasible solutions. The performance of BGAwEIS is confirmed by better quality degree of its optimal designations compared to algorithms proposed here and described in literature.
Weighted K-Nearest Neighbor Classification Algorithm Based on Genetic Algorithm
Xuesong Yan
2013-10-01
Full Text Available K-Nearest Neighbor (KNN is one of the most popular algorithms for data classification. Many researchers have found that the KNN algorithm accomplishes very good performance in their experiments on different datasets. The traditional KNN text classification algorithm has limitations: calculation complexity, the performance is solely dependent on the training set, and so on. To overcome these limitations, an improved version of KNN is proposed in this paper, we use genetic algorithm combined with weighted KNN to improve its classification performance. and the experiment results shown that our proposed algorithm outperforms the KNN with greater accuracy.
A Stereo-Vision Based Hazard-Detection Algorithm for Future Planetary Landers
Woicke, S.; Mooij, E.
2014-06-01
A hazard detection algorithm based on the stereo-vision principle is presented. A sensitivity analysis concerning the minimum baseline and the maximum altitude is discussed, based on which the limitations of this algorithm are investigated.
ZHENG Yu; CHEN Zhuang-zhuang; LI Ya-juan; DUAN Jian
2009-01-01
A novel automatic alignment algorithm of single mode fiber-waveguide based on improved genetic algorithm is proposed. The genetic searching is based on the dynamic crossover operator and the adaptive mutation operator to solve the premature convergence of simple genetic algorithm The improved genetic algorithm combines with hill-climbing method and pattern searching algorithm, to solve low precision of simple genetic algorithm in later searching. The simulation results indicate that the improved genetic algorithm can rise the alignment precision and reach the coupling loss of 0.01 dB when platform moves near 207 space points averagely.
A Coupled User Clustering Algorithm Based on Mixed Data for Web-Based Learning Systems
Ke Niu
2015-01-01
Full Text Available In traditional Web-based learning systems, due to insufficient learning behaviors analysis and personalized study guides, a few user clustering algorithms are introduced. While analyzing the behaviors with these algorithms, researchers generally focus on continuous data but easily neglect discrete data, each of which is generated from online learning actions. Moreover, there are implicit coupled interactions among the data but are frequently ignored in the introduced algorithms. Therefore, a mass of significant information which can positively affect clustering accuracy is neglected. To solve the above issues, we proposed a coupled user clustering algorithm for Wed-based learning systems by taking into account both discrete and continuous data, as well as intracoupled and intercoupled interactions of the data. The experiment result in this paper demonstrates the outperformance of the proposed algorithm.
A Hybrid Distributed Mutual Exclusion Algorithm for Cluster-Based Systems
Moharram Challenger
2013-01-01
Full Text Available Distributed mutual exclusion is a fundamental problem which arises in various systems such as grid computing, mobile ad hoc networks (MANETs, and distributed databases. Reducing key metrics like message count per any critical section (CS and delay between two CS entrances, which is known as synchronization delay, is a great challenge for this problem. Various algorithms use either permission-based or token-based protocols. Token-based algorithms offer better communication costs and synchronization delay. Raymond's and Suzuki-Kasami's algorithms are well-known token-based ones. Raymond's algorithm needs only O(log2(N messages per CS and Suzuki-Kasami's algorithm needs just one message delivery time between two CS entrances. Nevertheless, both algorithms are weak in the other metric, synchronization delay and message complexity correspondingly. In this work, a new hybrid algorithm is proposed which gains from powerful aspects of both algorithms. Raysuz's algorithm (the proposed algorithm uses a clustered graph and executes Suzuki-Kasami's algorithm intraclusters and Raymond's algorithm interclusters. This leads to have better message complexity than that of pure Suzuki-Kasami's algorithm and better synchronization delay than that of pure Raymond's algorithm, resulting in an overall efficient DMX algorithm pure algorithm.
Solving frictional contact problems by two aggregate-function-based algorithms
Suyan He; Hongwu Zhang; Xingsi Li; Ron Marshall
2005-01-01
Three dimensional frictional contact problems are formulated as linear complementarity problems based on the parametric variational principle. Two aggregate-functionbased algorithms for solving complementarity problems are proposed. One is called the self-adjusting interior point algorithm, the other is called the aggregate function smoothing algorithm. Numerical experiment shows the efficiency of the proposed two algorithms.
Ship Block Transportation Scheduling Problem Based on Greedy Algorithm
Chong Wang
2016-05-01
Full Text Available Ship block transportation problems are crucial issues to address in reducing the construction cost and improving the productivity of shipyards. Shipyards aim to maximize the workload balance of transporters with time constraint such that all blocks should be transported during the planning horizon. This process leads to three types of penalty time: empty transporter travel time, delay time, and tardy time. This study aims to minimize the sum of the penalty time. First, this study presents the problem of ship block transportation with the generalization of the block transportation restriction on the multi-type transporter. Second, the problem is transformed into the classical traveling salesman problem and assignment problem through a reasonable model simplification and by adding a virtual node to the proposed directed graph. Then, a heuristic algorithm based on greedy algorithm is proposed to assign blocks to available transporters and sequencing blocks for each transporter simultaneously. Finally, the numerical experiment method is used to validate the model, and its result shows that the proposed algorithm is effective in realizing the efficient use of the transporters in shipyards. Numerical simulation results demonstrate the promising application of the proposed method to efficiently improve the utilization of transporters and to reduce the cost of ship block logistics for shipyards.
Tele-Network Design Based on Queue Competition Algorithm
Huang Zhang-can; Wan Li-jun; Tang Tao; Chen Zheng-xu
2003-01-01
In this paper, we report research on how to design the tele-network. First of all, we defined the reliability of tele-network. According to the definition, we divide the whole reliability into two parts:the reliability of the mini-way and that of the whole system. Then we do algebra unintersection of the mini-way, deriving a function of reliability of tele-network. Also, we got a function of the cost of tele-network after analyzing the cost of arcs and points. Finally, we give a mathematical model to design a tele-network. For the algorithm, we define the distance of a network and adjacent area within certain boundaries . We present a new algorithm Queue Competition Algorithm(QCA)based on the adja cent area . The QCA correlates sequence of fitnesses in their fathergenerations with hunting zone of mutation and the number of individuals generated by mutation, making the stronger fitness in a small zone converge at a local extreme value, but the weaker one takes the advantage of lots of individuals and a big zone to hunt a new local extreme value. In this way, we get the overall extreme value. Numerical simulation shows that we can get the efficient hunting and exact solution by using QCA. The QCA efficient hunting and exact solution.
An Improved Dynamic Joint Resource Allocation Algorithm Based on SFR
Yibing Li
2016-04-01
Full Text Available Inter-cell interference (ICI is the main factor affecting system capacity and spectral efficiency. Effective spectrum resource management is an important and challenging issue for the design of wireless communication systems. The soft frequency reuse (SFR is regarded as an interesting approach to significantly eliminate ICI. However, the allocation of resource is fixed prior to system deployment in static SFR. To overcome this drawback, this paper adopts a distributed method and proposes an improved dynamic joint resource allocation algorithm (DJRA. The improved scheme adaptively adjusts resource allocation based on the real-time user distribution. DJRA first detects the edge-user distribution vector to determine the optimal scheme, which guarantees that all the users have available resources and the number of iterations is reduced. Then, the DJRA maximizes the throughput for each cell via optimizing resource and power allocation. Due to further eliminate interference, the sector partition method is used in the center region and in view of fairness among users, the novel approach adds the proportional fair algorithm at the end of DJRA. Simulation results show that the proposed algorithm outperforms previous approaches for improving the system capacity and cell edge user performance.
Robust kernel-based tracking algorithm with background contrasting
Rongli Liu; Zhongliang Jing
2012-01-01
The mean-shift algorithm has achieved considerable success in object tracking due to its simplicity and efficiency. Color histogram is a common feature in the description of an object. However, the kernel-based color histogram may not have the ability to discriminate the object from clutter background. To boost the discriminating ability of the feature, based on background contrasting, this letter presents an improved Bhattacharyya similarity metric for mean-shift tracking. Experiments show that the proposed tracker is more robust in relation to background clutter.%The mean-shift algorithm has achieved considerable success in object tracking due to its simplicity and efficiency.Color histogram is a common feature in the description of an object.However,the kernel-based color histogram may not have the ability to discriminate the object from clutter background.To boost the discriminating ability of the feature,based on background contrasting,this letter presents an improved Bhattacharyya similarity metric for mean-shift tracking.Experiments show that the proposed tracker is more robust in relation to background clutter.
Beam Pattern Synthesis Based on Hybrid Optimization Algorithm
YU Yan-li; WANG Ying-min; LI Lei
2010-01-01
As conventional methods for beam pattern synthesis can not always obtain the desired optimum pattern for the arbitrary underwater acoustic sensor arrays, a hybrid numerical synthesis method based on adaptive principle and genetic algorithm was presented in this paper. First, based on the adaptive theory, a given array was supposed as an adaptive array and its sidelobes were reduced by assigning a number of interference signals in the sidelobe region. An initial beam pattern was obtained after several iterations and adjustments of the interference intensity, and based on its parameters, a desired pattern was created. Then, an objective function based on the difference between the designed and desired patterns can be constructed. The pattern can be optimized by using the genetic algorithm to minimize the objective function. A design example for a double-circular array demonstrates the effectiveness of this method. Compared with the approaches existing before, the proposed method can reduce the sidelobe effectively and achieve less synthesis magnitude error in the mainlobe.The method can search for optimum attainable pattern for the specific elements if the desired pattern can not be found.
Performance of Buchberger's Improved Algorithm using Prime Based Ordering
Horan, Peter
2009-01-01
Prime-based ordering which is proved to be admissible, is the encoding of indeterminates in power-products with prime numbers and ordering them by using the natural number order. Using Eiffel, four versions of Buchberger's improved algorithm for obtaining Groebner Bases have been developed: two total degree versions, representing power products as strings and the other two as integers based on prime-based ordering. The versions are further distinguished by implementing coefficients as 64-bit integers and as multiple-precision integers. By using primebased power product coding, iterative or recursive operations on power products are replaced with integer operations. It is found that on a series of example polynomial sets, significant reductions in computation time of 30% or more are almost always obtained.
Visual tracking method based on cuckoo search algorithm
Gao, Ming-Liang; Yin, Li-Ju; Zou, Guo-Feng; Li, Hai-Tao; Liu, Wei
2015-07-01
Cuckoo search (CS) is a new meta-heuristic optimization algorithm that is based on the obligate brood parasitic behavior of some cuckoo species in combination with the Lévy flight behavior of some birds and fruit flies. It has been found to be efficient in solving global optimization problems. An application of CS is presented to solve the visual tracking problem. The relationship between optimization and visual tracking is comparatively studied and the parameters' sensitivity and adjustment of CS in the tracking system are experimentally studied. To demonstrate the tracking ability of a CS-based tracker, a comparative study of tracking accuracy and speed of the CS-based tracker with six "state-of-art" trackers, namely, particle filter, meanshift, PSO, ensemble tracker, fragments tracker, and compressive tracker are presented. Comparative results show that the CS-based tracker outperforms the other trackers.
[A new algorithm for NIR modeling based on manifold learning].
Hong, Ming-Jian; Wen, Zhi-Yu; Zhang, Xiao-Hong; Wen, Quan
2009-07-01
Manifold learning is a new kind of algorithm originating from the field of machine learning to find the intrinsic dimensionality of numerous and complex data and to extract most important information from the raw data to develop a regression or classification model. The basic assumption of the manifold learning is that the high-dimensional data measured from the same object using some devices must reside on a manifold with much lower dimensions determined by a few properties of the object. While NIR spectra are characterized by their high dimensions and complicated band assignment, the authors may assume that the NIR spectra of the same kind of substances with different chemical concentrations should reside on a manifold with much lower dimensions determined by the concentrations, according to the above assumption. As one of the best known algorithms of manifold learning, locally linear embedding (LLE) further assumes that the underlying manifold is locally linear. So, every data point in the manifold should be a linear combination of its neighbors. Based on the above assumptions, the present paper proposes a new algorithm named least square locally weighted regression (LS-LWR), which is a kind of LWR with weights determined by the least squares instead of a predefined function. Then, the NIR spectra of glucose solutions with various concentrations are measured using a NIR spectrometer and LS-LWR is verified by predicting the concentrations of glucose solutions quantitatively. Compared with the existing algorithms such as principal component regression (PCR) and partial least squares regression (PLSR), the LS-LWR has better predictability measured by the standard error of prediction (SEP) and generates an elegant model with good stability and efficiency.
Novel density-based and hierarchical density-based clustering algorithms for uncertain data.
Zhang, Xianchao; Liu, Han; Zhang, Xiaotong
2017-09-01
Uncertain data has posed a great challenge to traditional clustering algorithms. Recently, several algorithms have been proposed for clustering uncertain data, and among them density-based techniques seem promising for handling data uncertainty. However, some issues like losing uncertain information, high time complexity and nonadaptive threshold have not been addressed well in the previous density-based algorithm FDBSCAN and hierarchical density-based algorithm FOPTICS. In this paper, we firstly propose a novel density-based algorithm PDBSCAN, which improves the previous FDBSCAN from the following aspects: (1) it employs a more accurate method to compute the probability that the distance between two uncertain objects is less than or equal to a boundary value, instead of the sampling-based method in FDBSCAN; (2) it introduces new definitions of probability neighborhood, support degree, core object probability, direct reachability probability, thus reducing the complexity and solving the issue of nonadaptive threshold (for core object judgement) in FDBSCAN. Then, we modify the algorithm PDBSCAN to an improved version (PDBSCANi), by using a better cluster assignment strategy to ensure that every object will be assigned to the most appropriate cluster, thus solving the issue of nonadaptive threshold (for direct density reachability judgement) in FDBSCAN. Furthermore, as PDBSCAN and PDBSCANi have difficulties for clustering uncertain data with non-uniform cluster density, we propose a novel hierarchical density-based algorithm POPTICS by extending the definitions of PDBSCAN, adding new definitions of fuzzy core distance and fuzzy reachability distance, and employing a new clustering framework. POPTICS can reveal the cluster structures of the datasets with different local densities in different regions better than PDBSCAN and PDBSCANi, and it addresses the issues in FOPTICS. Experimental results demonstrate the superiority of our proposed algorithms over the existing
WEB SERVICE SELECTION ALGORITHM BASED ON PRINCIPAL COMPONENT ANALYSIS
Kang Guosheng; Liu Jianxun; Tang Mingdong; Cao Buqing
2013-01-01
Existing Web service selection approaches usually assume that preferences of users have been provided in a quantitative form by users.However,due to the subjectivity and vagueness of preferences,it may be impractical for users to specify quantitative and exact preferences.Moreover,due to that Quality of Service (QoS) attributes are often interrelated,existing Web service selection approaches which employ weighted summation of QoS attribute values to compute the overall QoS of Web services may produce inaccurate results,since they do not take correlations among QoS attributes into account.To resolve these problems,a Web service selection framework considering user's preference priority is proposed,which incorporates a searching mechanism with QoS range setting to identify services satisfying the user's QoS constraints.With the identified service candidates,based on the idea of Principal Component Analysis (PCA),an algorithm of Web service selection named PCAoWSS (Web Service Selection based on PCA) is proposed,which can eliminate the correlations among QoS attributes and compute the overall QoS of Web services accurately.After computing the overall QoS for each service,the algorithm ranks the Web service candidates based on their overall QoS and recommends services with top QoS values to users.Finally,the effectiveness and feasibility of our approach are validated by experiments,i.e.the selected Web service by our approach is given high average evaluation than other ones by users and the time cost of PCA-WSS algorithm is not affected acutely by the number of service candidates.
Research on palmprint identification method based on quantum algorithms.
Li, Hui; Zhang, Zhanzhan
2014-01-01
Quantum image recognition is a technology by using quantum algorithm to process the image information. It can obtain better effect than classical algorithm. In this paper, four different quantum algorithms are used in the three stages of palmprint recognition. First, quantum adaptive median filtering algorithm is presented in palmprint filtering processing. Quantum filtering algorithm can get a better filtering result than classical algorithm through the comparison. Next, quantum Fourier transform (QFT) is used to extract pattern features by only one operation due to quantum parallelism. The proposed algorithm exhibits an exponential speed-up compared with discrete Fourier transform in the feature extraction. Finally, quantum set operations and Grover algorithm are used in palmprint matching. According to the experimental results, quantum algorithm only needs to apply square of N operations to find out the target palmprint, but the traditional method needs N times of calculation. At the same time, the matching accuracy of quantum algorithm is almost 100%.
Research on Palmprint Identification Method Based on Quantum Algorithms
Hui Li
2014-01-01
Full Text Available Quantum image recognition is a technology by using quantum algorithm to process the image information. It can obtain better effect than classical algorithm. In this paper, four different quantum algorithms are used in the three stages of palmprint recognition. First, quantum adaptive median filtering algorithm is presented in palmprint filtering processing. Quantum filtering algorithm can get a better filtering result than classical algorithm through the comparison. Next, quantum Fourier transform (QFT is used to extract pattern features by only one operation due to quantum parallelism. The proposed algorithm exhibits an exponential speed-up compared with discrete Fourier transform in the feature extraction. Finally, quantum set operations and Grover algorithm are used in palmprint matching. According to the experimental results, quantum algorithm only needs to apply square of N operations to find out the target palmprint, but the traditional method needs N times of calculation. At the same time, the matching accuracy of quantum algorithm is almost 100%.
Scheduling Algorithms Based on Weakly Hard Real-Time Constraints
TU Gang (涂刚); YANG FuMin (阳富民); LU YanSheng (卢炎生)
2003-01-01
The problem of scheduling weakly hard real-time tasks is addressed in this paper.The paper first analyzes the characters ofμ-pattern and weakly hard real-time constraints, then,presents two scheduling algorithms, Meet Any Algorithm and Meet Row Algorithm, for weakly hard real-time systems. Different from traditional algorithms used to guarantee deadlines, Meet Any Algorithm and Meet Row Algorithm can guarantee both deadlines and constraints. Meet Any Algorithm and Meet Row Algorithm try to find out the probabilities of tasks breaking constraints and increase task's priority in advance, but not till the last moment. Simulation results show that these two algorithms are better than other scheduling algorithms dealing with constraints and can largely decrease worst-case computation time of real-time tasks.
Implementation of pattern recognition algorithm based on RBF neural network
Bouchoux, Sophie; Brost, Vincent; Yang, Fan; Grapin, Jean Claude; Paindavoine, Michel
2002-12-01
In this paper, we present implementations of a pattern recognition algorithm which uses a RBF (Radial Basis Function) neural network. Our aim is to elaborate a quite efficient system which realizes real time faces tracking and identity verification in natural video sequences. Hardware implementations have been realized on an embedded system developed by our laboratory. This system is based on a DSP (Digital Signal Processor) TMS320C6x. The optimization of implementations allow us to obtain a processing speed of 4.8 images (240x320 pixels) per second with a correct rate of 95% of faces tracking and identity verification.